var/home/core/zuul-output/0000755000175000017500000000000015150370327014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015150373416015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000204052415150373252020260 0ustar corecoreikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9Gf܅ "mv?_eGbuuțx{w7ݭ7֫'^|1Fr_?c^*߶E٬:rv筼ح_y~̎+\/_p/Bj^ֻ]Eo^O/(_/V?,<']_kmN:`Si{ C2i1Gdē _%Kٻւ(Ĩ$#TLX h~lys%v6:SFA֗f΀QՇ2Kݙ$ӎ;IXN :7sL0x.`6)ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&ccNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4V{g6R/wD_tՄ.F+HP'AE; J j"b~PO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTt\EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL8G9?{WƱPz;| \;_D[T/BI GH8@"t*"9z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉\^9NPHGSd?'xc 73\+s$N9tf{w5n%՘#TI{h5pfIWGev.F_{!7*qW}ӼتY&7!5Y'3KtpwU`<d5J{@ݜ P?&pWt|CX6|,9:N\8|ҳ uE UIL`F &ni?4~j3?m-Kھnۅm''J[/m$g4rfrvRޟ/O_·ogg]fyrd-i-Iv#GL`,Oȃ1F\$' )䉳yg=#6c+#  =J`xV,)ޖ,3~JPͪm|$oV1yU<̐t6 T m^ [IgINJ\Оf*Z"I)+>n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|\΋Z " ;9;/f֚Znϻ-Rp^Tku$-#ہ=ll[/؝nA(ȱbAj!%YdE6c>!Ƶ *0W4Q>@>lWN"A X5G-nm.8B.OI[31,j2Ce |M>8l WIf|\q4|UjA.gr`˱LO{ xr.~l-ɩu_DժmbG~MP%] x56ҏMF9 E"A,c V}\"X.2< 5FB΢u-`aL#ʄ%EGƶ*| 흰:S PxbBOίr^3>Uf1w[mCjaZ-1_kٚ%Vb\͙#G6 `q+MPU~m!.?IWPĽ"] rT [Us؟˰ ]\ h! n˱>5S1px k}sRmA>d2U^kؖlX܇Bz1U_#rTE6u$c2l$5|vg>=u^K+Ⱥev/#miX Z*ǪCGJM->=?! öG6! K><Q)D 1/y? bO]?}!vy3ޯMPu>o>$lO۷&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz -e7d_^)3 נׂ6-CokIf#[s,R-&)2$?vlW$,11'ˠ*.q/E b\ix)6tb"H~%.b@:Kec6NZV8FDza5H`:&Q5 ^X5X-&$&*a6>B>+q"/[ڲ&Lh!*<-ol0qiԧ {E#A7eVG&EV$_gWb ۞. E[@'nf"db'Z q3)|x!mLaKfKyࢷgb񾍠z}(6>C,HI~'.Op% 8$ c*Dp*Cj|>z G` |]e*&q!t=|X!~ Pu(ǍUeS@%NբZ.SVw Ohgh uЍaRs ^d6GX^V;/+ sXju~NF^)!FLѲ?VK.qE╯YOȱ<~qWP~ RT4&+QR"tJ8ۭSOfbxRKz .[c&Mȩ f(M`,mMr1E L"8_SK$_#O;V taUێ]<&YnrꝤʇ)br\kB`X}nl}hSЕ΀ѩ آUzn޷(ȊD0N^`MDN74ТC>F-|$AZPB8dJU&4Իflq6TX)ى?wg6>r\5TT%~am.>!LcoJrKmzqvmz܅EAZ#u-9`x 92$4_!9WՠZ̓?Wnm>0Es%DƖ|2H\2+AaTaBˮ}L@dr_WԦc>IdA Od[=jlek=XJ|&+-T1n8TڎP$%s,qgt+ZSxToE7U9/nq.JY):Y:7AIU"cKӝ$'qo%\Q!%c5\Z9N4Zxz,dI*ƙ(EfE"`{ipEIՒ9| Olz3;QϢ*:]ք+I&s5w` q:CdʰH`X?"}B=-/M~C>''1RWX%2@KʸH'sۄ`gRpcf:|XUZ#OSt/G~-~o2:u)"\**vdC_ˆdvcƕMlA&HwlF@ա5+F>ΰ-q>0*Oѹ eO/I!m|xV&\b<9$4Nvm^آ]$GBoMjKٕy{H 31Հm-PġUX$[eR6Oœ-m~)-&!883\6y 8V p-lprG]斾-3jsqY~ sj\+9[rAJsT=~#0t2ެf¸DŽ,@2,?WYخNr<V` =V[B5!Z\ļǪ:0A*Iucv8\[|۶s L-ky{ K?_a2=c5%C\d\'2J1(Rve:<+A/VRy6 ö+ML-yz,ZlQ^oAnv-{)xǺ--pcl@Ֆ*Vߓ`ڄ(Nc읠}*وGmU`pi|5Ӄ~ &$yx `qJA"*1 [CplmWu'T0^!Yg5;о߾/5I7wfܛT'W_};Gsݸ]qY7߃뭽*{Q=^]{Q_Trin`jSت-ȥ{Q=\;ju[}.˻Z]=F<^[UsywaoőPi{4o0ޮ$WOjbW\cE~yGd }va&C`Jꡝ/f_Œ8bJ*xMY(=ίl~l8V0t٪T z{Ac:&$ NƛCpH*DtWdT)*7U6IXabcuyVI>UlkO|Ty$ŋrOV&nIdE Ͻq*nŘʰN҂V_޵6#ۿ%r:LƵq;TTkB5odɱZM'l$ztn扼$m,yvIOd^S8T#4_ fPpd Jq;,J.C38TC, X˃_FEˠ.E9oۓxUOyH#j0n!K xRzGupF,@2C5۰\{!ˊiEB8,u%wB\Ao>PG8g"'Yɂ8<'QpBOtG3ƅEZpeYnW ~{;I~s3.)ʶ]Rz ]Rzqj +ceFF VΞzhf,Ɩܗ%cv\ǖ~獀z˒"cxYRzl~gm(L]Zc9n1 gfRʫ)Ve5tǏ}P?gޫ5)!>7dX֫r9:y?QO0l/= v0h'n^d,xA,},/|~Av%k13-lrK>#HDKO!nj/eyH}]f$;}HNfE!}IE:㽳[b(ۈ 螯AhjCuIFo7tg$ ew-D[:W\WNΩ^2 a8RۂtE>^TaeE[E@VbiiU2 !pd\J>dg,YtF+ ~7pϸAX1T9 roM9<.yIca.9[f4VߖP6z[u1h!y /NGK Bpvd5.5&閒6<{?QAW8e>3*@?fe=-YIJR v%֦|֜oh%yCtYb-ɇS@]݀ `l6U( ϖ0][}HJ8'iv}]ь6eV9>Cqe$ǤЮAOuW56gINf@5~vٜo:m\,Coh)cBzC!GȏVF(%A(w@"/*k^OepR~ hEJB:Kk^~eUHn[w:_/A3߽"G2j8"yO#2d8:>dqi4'dJA6H`,O*O U.v?fЙw'o_4F#}= :4f iX%'$l䣀G7O$/eJK!G>>:InɀnpmWhTAЪ0}>hĦU6%0ikVH8{ZԬGi Q8>BBy}EoxŃ W( l-:V&r>XETVƲZ{wV4WN|ϑ > ZC,<=2OuB4Zpww7\57O" d6?W+ Hx{fOsדPu/-) aD=Y=eqEFfEs!BuPǢXhQC O 1ϣ&'JW9$HqMTT-oHO[zچ‘^er[~yV$Z_C`Bis*i.|.7@϶ }t92dZuS Y-lDWL/uPbŠO>h  < XGiTX|)aXg=d &Ԟ%RT]:KCڌ{rwBmgsü FO$Ց#Uag4ANZE KadM=Z2Rt{vV- p]g&,oܕf_AxFO^$ˀFz.%Z?&]w 1 $zZsZY6x+!-hȫEtNp=-1Trh $_6Nj1ߌ2p=5ד)v 'TY{- Vűb2vOeVsSuo/W ,Ppi{X\{wFA-6>@rTh .B+Y$2` fԮ[_ϧ-M$;[u-Gz%[Q 5hp˓EU wZ%|gW2ғ..lzrmO;>".,AJ>i;n!84Q=DgE^W$YabNupQ';[#v>ڇV5<}V V{H`IlQ$=G@UȼGBtH.;x۴ܝ▹4d|3Ӗ9x{s0!߷quk/Z7GG\گ#h{'zC_M0{V"g}|=e~;w41ғ ö]2wTޞ\M\ܗKO\CQzk0dzjE\D+H^z \78on(̢1TC佃usi$Z7i ؟0@|idnYM#l&OY|f_ܺQ>.^  *2?{6 U;uwkJ>|S1\$HD"uǹ߯ )aGI&H4 irpFw&D*e]Q]F'4b&КӢy[}Bu=uzX/Y #xHzUG:sUe?.yIx~Ӆ\ $@Pa8pD %x?!dRVoC7kD~ݕtVR/,ySqBWC@+Ʌ:L#ӞBIV} r2Tӛ[q(]JÅx{q˳;`youƴ`Ns+L0to]QY9{g `8st+>%ȑ2H @d`@)]F>{r瀁w(0phX8Rs!hjXP:2/eLIt7+' N)9c^'"W8y>33-p@ԑECݛP<OGLqaܯ/@q<ځ6t$t+ "xd (?E^(n[s&#HnSc1H n3CtݡP{pU,O˯_K•MJSP MX9'(+ɰ,Xn4azٛcܽAab0B͔0Wc"0vi 3dnÑAA)t4k3%# Xk}h+y]W_RÐavd N"2Y,thMA@ьBaOqN|F. ieĭ:,5|4#"Uvu܋)*dTFQSEXJ}+CrVth\O* XU_Ͼ.\WdX4"xBlG-JCĝF&>u\E}b4tA&01S^G, - rCI)} h< РYZ{t,sOVf&$B=Jfx:NUsZšQ݃3N4zVa'ȭWh|ՓN^w$̘݄;rME~2ݔ9^OқM3i`% TE_< gIRX߽e&:&MN{-n v; |ts~z>qc/8asJUEYABe'Ê @E ,ք[氺"'j`{R oj[kyL (V)l[WZ?!X\XKVYd?^`O*k, SP,>;j5ጻsns[%g~O(O^G3 8l|QW-g,n^(p<]!-i:.{<AH)1"|cPyh2yP<]~suVܢQ5꯳ )T`+.ꭖ"F} !u>'xyq !Phϳt`z]ӀAU)M#V[Ba4;,`6 8uG[Ù8I0v:+NizkDOe]w+M WmhQ=N5eu'[WOF-}LmīeA6GG}O << ȼdc ],r"Nl^üȱ&U$fz(IЍ.F74(n(: QBl=iCX]?KO]|}w`c=>FkJ{ ,n*nrt6sܓJi0mȐ7iYL ZZvfѵx_B oh5=9VQp{w{rtFY益9tєD&eVM*p*s,F-1-Ab?N+ǑʅsP^l%&i6V\<)O"c6<}?8'~~?ʬ@C-W - E`tkASGMۂ4s,RMX&DCpu9\:S sS=;Mbsu2 R J)DPEAI~?|0FhNs V_QWI.xB?i| ͯ{wz( =R @ ݅!N˚~l-%paZvt2 -<9_h6> SaLd+sZ$\u|r`̇EJ0gLDجh9| ̠-fYch/\㴌F&[cfpo,b ſĕ a>") ˧zI&QWF3A?f1Abx,hu)\ :0N0D#;D6Ne@LlB VL;= Fn [ т MĬyT$gYRH6nn+. &pBD`\!vUko/Kʳzga~VgS_;W \S]qHP.VHwf,Ce"$)yVgpGE*A$'PgiCDz~*Т!^i[#uǝ7g𥳐bt+}'zulAuؖû>X3byr7F׸W0,kZ7||~VAZQWM^ ){ [pVM$D4:8Gpܦ&A5`Jz/b;t#3FVm, (L0%DM?>{{L'"Lڈjj]$F)V`Ҹʪ($m~ R$6aeQ&$lRhJMH{hi ^w:S pY{[MT$>j%mҐkїV3#r|hX=!E۸6jb)0+ɩGs?Es?-NHr^eQ;jT1]*`΅82-D"7}qI!VKBp9n msDI6Y_1g׾P~ヲ&TtQg K^s FWmE2!=1< ,8.ghr6BAܗS;悰V+ G s|ʂSqt-9Rm4Z彲 #H#{dwxctiVStA&t6$#zc7Ăݿc's-k& 8 ٻ)M6Gd(0Z}'eu*XcВN^4{9೉՟a60" ޶tsxrAa5~__HH_daہŏ(5׏S^W'gT}'zҝ >c\C謈- ^Ƙ C4zDOLܢZ 3ob+08Ҩ{يmRZ$D1zVfndJKAe.DLj-2gQXw!Hxհ9KLDŽFu3L^-.@,2P$-e&gb3~ee" }9SUxI0ӹ arbWU ]@Q(ӣrB01d@M6r F oO^bumubdEH!%JCOx@;99 (L 3\g8~ Չ$]8YpTsvkUP*Q -jA$f'\,UFo mKjR!4%HO KL3`bX!!DLּJL۫Bs^`M9Z1W*A<-SC_GŚV#iQ$q7ybڪֻ#U>ZXpTaUyꑒ`ng\%M͜F9ykͪwH9\V782I2\N卙 t)7d2X?:Y=kݦF,P|l"zuH(z! R$]U|!n\{R{bX iEM[\YEH-c r2,}U)x7'/EmZX7&)NtDخ$Hg>fՑ|T' bh@%G0.:s5R}#Eܕz&u?8yKMdk28Rv}|Dy{IG?;_VmoPwg4nrNi)mWvpv|1UT,)<=nWYpsqv9ژnM @E'1:q[bo"ǽG#_>ed*e@S uYps3FTLr D_+#E3 fn"dʺV AXTTS%YY['elJuTQá^Xpt3ڧ1u9#|X՘{Mlr z|g;vHZc69yKڹEB[~Kl-*8R*YTs%chc(G8"Ő8.΄"ki2iZ*WU_85P O8>8ݷ5i6IKo,>Ǽ7*YR+)5I3,\Ƥ#^ _;v ez&V*K sO8UV0Hu;~,DQ* J& }%rBnǡ2yÌ/Dr]yv p z']:owS6),M'2%gP;^GWۡ؆旀~`>D䴽f- ,&@#9N(yJEm7U TJ(pd۳&)SIz61:SAĜ,oPc*mT0 ?uNϜ*qLQ^s 8F(c[:وIf:7dܵ_sJ%&tӲ0.Asp b0Tk4B]&Zh:7ER"qܬ9ΛO}aNv]: [v1(;\u>!aWcܲ"|>=>Ī0˨ЊYݖrܖle>ݑky\v掽CcKtA%ʂQ+qD9/tzz͡^L|{g:̒= ._̧r+KTEHwqN[}@c?VB 8!MNUOL).vWXp۷* ̷v-OexpeҮr–p88(qjSOj:$<'Wk.(ҏz?8֤ ytiVC̚$f*E.;%%E w+.w$QE\\qYp?2vmoj?l?ܱhŔE@ݠd]ʮ`IDp# ;R-(TDdNN&ถy-7gJܦL9z 8 ES.&,H}>ă_{cANq1~/,8:Wytd<:TB|;e&WFen/ N J^~lw?l9Kvk)BNGo1MajZ.n$疿><ܲv2[ NfN. VWV΢zdbwt-2 p3Yp|XFgg|)BASE㉣I8?[.70}mELPC)'<*.@l?źNȼ<ȔS]/ɷBeuRv5%O΋4)XoF58AS:ܛȂ[2,p0\2 9{I)?LcTg: ܺFptK#.B>1DXxLBz|%Io8n7V BoøhSq[5!8530|9{Z9d, nK|?$OptCx ?AdFXjp)D''chJlZq7]\|O΋n`DiLgdti̾G%`4l[=sx4x|wr& 10ǝ;z|3)ͧU|[)f09:.3L$-48B8ariǣ^*eswUg("ˣhҕWUh(q pO '|S,@drY bCbcM_Gg}G3]j'  G+!؅_@  㤘˂CȡJWm''?%EӲM4r`+&śzHZ@kh?;۵m?> 3Z5XIa^49*:LC1⸨Sw.\ |[N]7'B46ZGx]7 aNZqDAJ,@O):8]T#c>*~¾!,ݙP됦O8κdF'`}/ɇ5$:'N0i sֿmD]ͶȨ3C6,;u]r* fYޛě%}w $it~On4\[GLPjo .. ĀyO(J\ؾiYVbΌ;fr~;dG=QFy:ppMqlCZ(\Lz9Wgf7{oj t?[MvHM(~/ \/.̌E,_QqmHn :6#UaD߿~+by9q0Ty00rRzm洬^[tyf?zQA:~﾿#ҍEuo]W\nFLyݬDډ+w\7k%-G" MT$ҌƎpKug)xj SPOt.# UHHFT!%PT`VSj&1ZְDZvUs< 0hi4,kivu @ ֙y UVܠ5Ɇ$yYLmX+xF@T%ԏZַ<C8bYF;_N\OZbHhpiNO"is~\('B9Yto: 0HS`|i`P? qHKYmOQ6Q qi\()pPu{Tr4`TvZNYd'=EE/A Y2Y$ƹVҦdOlᶜ'#\2OкZ(^boU||dMIG府TԍRMP)S7V%7@onx9#ٻU l2{r|093H}_&Z^jS2-pc4Ȗ}* ڕBRl ԖTR|FޒF^a[Ƚg^LFed[!݀mϯlvY8*h]05p dB6g IҪC6~[Gf:p~&+~ųS鞖Je|6_j5haM_ !w-w Ig}u;}nق3LbEt[4λ]H7GTڡh ^w)dZ 632p8BbDQQQ۵[BLҽUFKXDZ„{Q*@kL E3# ie5 &&ֶuw >T{7b2i 3adHi1 ~QԐ(=:(˜)Vk[F׎:YFnZ=UsK1 pRpze>hQ%&@ rScu;v,c3߯Yuۯp%?I4ĕ*HV(%?Io\K~7|"hz/\ C0.•W t\k V>8"Ȕ]wCE:)/;ti!/;ȝ/W%R| z8wC }ª f酳#%W.3$Oհ=4^gjr\) U(M||D>MlY\N-@~`\ˣ@6nD& QUAZ`Afi|=qU;w ?MFa/ Zߞ7ES|5t?F|7dziVHziUo{)9+_浙kဦS؆w|+*׵ >62TkAձ0 *"T'\5ո]':I)k[-^+\@c u$ B"eU7b01T~t?onᨘxP^?>&ێZ$(8o({ F5Y:cx-Ra AM7@Wь?|?"IyVXXNy40 UsX`Q 8ra@+΃#Xh3-wl&ш EY-FY:+.5ewi;|FمV.;Z;&1!5 l sF'1L sUHz)Pgʚ ewy[ΊmDK:)YQ SrT 5hAC^ru!s DjGET*o]@j)@m*.;Πvq]--Ԋ_;1W+-NQyx*Ar3_KɭdM#($eFxbb9& 3)@6mIZ <G琸{Zv`N'0L P,2 xS*$vRʧdQG"!5$2x@ ƐdXƝh0#S-ECm$GC~h|Pi98>K:BʣcZ :##$c@5@ht|g-|siU bKNp3Ω8"<RGh 8(c!Ä_!‚ѐ; cèW,5]8o"^b=b"")D \M0A!laHT3 Nh3S&/#Jp! *MFRAQ4@5)Z=>#$)Έu2EqfeiQFq$Rh1េ6 ᨩvШt@@`5 ̍2Er$rʈH.rb*L40+P>LEYYEgd  벴sm3ۂ9_0mH{[?1ebe;v0䞓fCޕqdC`6F]zwcl&c(DI)G%jvǖf}:޳ ZZ[Vc[x6mJ i]8"WS&LEűf g0HU(*V=0ki(M,A|p?0( ւW~]%k9(ZłuBYPiRC>P,dB"DDqg,)W~Uy,ZW1trP;VnZIVOFQ:Ytl2~CUw]{$" sdZvTZM]QYK2H}NAAզ6рSW ̎jG/i-jꓹ8aJ+{{D71 1"2g$K2f)tD$br.emI;2FME=6ƒN4f6A,Cl/B 6,$6;n1jݨ l<2AQ{ H"l S*ZHJh*T+cb@dRd4Ddq,NS|qe& ϒ8*#e8Km( !j*  ^gEv/^ N׀咇vLߍ;A^W4ή?%!52_D`sNjCk %=w}D Cq, G4*VDı%+t3!cpA8}̤V?l1c] ):TjY͛OT Ba$W'i_?WsuL!Vֈ)`Dmš[ubS2qg fR(nQmS,*ͯSw77Cta 7m̛~㕭֕=vwxF-h^Օ\/3 3^ ^3R3ѳe kA*N}iVA Jj0kp7:*LʊjJ/,e?~檊SX+cɰ/b)J6LVt|L/rzyjBɈ@{d__dpoC%}^o٥~:?<]vzNȸI9` hWk-|m7q7ҍ6w)rQR2x1C{n/kn{h]KmYޢb(CĽz[Ka5uaجf)؃ЖQK+}g$= dgT)C v3#Աȭh:r)yMՆxnA<ça(>_+:adT>b_l8{o۴"S5}\q1prSciGgA0'<$+[zq$Գ`{UE'?n 5V~\JQ#Fw1S굀Mt&^'tU|Iw٥"6\K[1>/1K> }<S[ {>|n2UXڥâ]i?Wn.@D1rOo쿟sʂg[KF3PҤIF! MKb]? C9&ş/ooG4ɿPu;Mv|^|hپ][/_>8ή_p/SwW/,h/޻RQ?݌ngS,_3,W( n׋B ݼ3 a5ݡpaObMnᅫ-h] i{wb^!\C^..c[4{/oNy@c@(4-UYr)P!מL uLJ49g]Jjj% C2.bi8{{z-= g>hGnmPQP9nVk{;]^udO=zFOd.nǬ#%oÇw'ᰘ?U%N潕ϝ:}{\t{܇XJ?{8eĎ& I~#Zq ގn?݌p4v'5>U˚8!g_/}p`c]ҧk)Wleڼb^ؾ{5%Gy h!sKH:s/LQNS&]/ax>U-7/xLTȅ>E?r^i֒֔3Ę%-lxH a{Vu!Ȱfa;v:_ı9}~~Ϳ[N6/_yzB(Crz>~pU@*A MgiU9=XozxJ,|*KD;10Qo?uW5󳶘R'K*?BDjEE!O;&q'e(*y*^P҇n1E%K /U]0.xg٫eϯ nۀ1E2IQɼ .!F+Ue-gLԱڙ긼v6:C vXq&i4]lVHh^ةf{,ъG\L!p?8p_UִU\E6{)NW mMtͩG89ΣW7d1J3`3DJ+ ΁p5]RWZUqӺ]95Cp"t~ұdB2p ɨ<τdP2 ւRc4c|,e*^~1Rjg8'=z$x Aɳ U#xY6g.PxYk~k%lތݰ sHr@! mݖ o2/:dt7eoc Ǵ\ZUX9+W4 vS65UhyL͆WY*_=Ƨ nmGYH޺rhF=:7Z=[ohɴn(ٮ@83HZeUbjF¼d; =*Xaa L Vlb[i{.k$Ԯg\ o{X/[h8t۽ˍaB'WžvpҒy (p#m1I^!]LRmig[H 'u8\-C{;T}5|4P[\dZ2Hl#ܦ:F+M@Gxvm@vFTr9)>yq.u &LZIp`dž݁MjD(MZ8D H@2BYU@ήH, BU\Sy3! B{k<+}NU DQ|nx9&AM u :J[F*00(S0Fq"vXԍ6Bi5>BRA N(O l44m2,څ%8lH`1aG8[b-au*\4 amN6T2}KD1Iv@6uQv!RsQ{kSܝVKQO=5m0I_E=ۅ*9O Xr@/RxygWi[ Wl$P`oT0g:'o V}{N5"R"3sZS(;f7*L313v6g@K#:CtdM 鍚P?mμ9)_Zis͙H3@|a4)/%? Dj(J_f VrW5l*(PЈh;)5|4 jʽH8,4]Hk{8)2 }Q:@ɪR8PɞO(P 5iŬV!JUdJYyͳ퉗mp@źՌv=<2!?a "U:w@l5 U` qb (Tn|Md~2GP`Ah"R\Z[SwC\5݄k:J9j9B d̪+@'Њ"!BD: "E*K"[o "!^h bvb .7A*CJRv n BJC3%Bɕ#F@G(y(LX eR%z%!_N&cG 鍒k@mOVX rŲ)ڸ=?;G+8Evb;S) =]dS1%B B'SŽ*B#G@sŠ\!H,"Q8+EvJCj,!'11+r(t *R")r7_6Y˽QBI.P*LPVgwa"Z׌"BFpCw{xPe%+TxU,NTYgM>Jc!QakH 0(MTx7i20idFۈSQBzVy@( 6%2g]4r+MRcGJS؁<].eHFLR`YKoBֻ#d2k)=.1L>e'q1֨HIrFL4I"ڝy@'wj~iOG||# y_~ ?V\oBuin<땿Y"?}!׿/jєNJ؝Cnp@Q^~N,1nu1c! -'c6Hfa7W!g@ڊ{&|9 LlK,"n,wyȓ*{48gښGx/&weh̟'`Xludӟj\´ߧV ?LO~˓h2jG߷_kIrcUﶞ{zj 2׃uZ0q*Jw8XҘݻLо}Jh$=ծ(g q0RYq ąTdT<5ON/;t5x1X֠jv`0y}`u񆡹wg,l\3FD谷x,9}oy è]Ƿ_Z?r.@i">\ uGuataZf@Mvcwp8?|x pdlx#i4~&uf*;[5y85u^iFZ õJ1C l-R'UPmHi:l2|3wVx%yL!rTZJ]XuFBUΉ.^~>72ϗn,jp=DmUUnj Ap=Ix{~^mZ8 uzFd&k SVz-Y[[~Y[s +eT] OZѸE7Z<3^3SW:,yBdu6$\[}he1>hsu!f g5 gmDňH)QGҚrJw؎Z]u}Q{}YGۢ,RdHX2Ʒ^!{h?[Z;tss_!*Wf MPmf4,J){"adV.z.dTuLHa/a i0 #txEN  SN{w͊Ex:g QweT 뢃9x~;)9 _Рmڳ9M\ {Ӌf\a#?۷}ulwPF(o62fnv;>Cm V{mqOsߌy l`wpg+B{km#I``~ X$]ap9 @C?ʒ#Ju=(NSwXSuuuW=U0OLq5E<-\Hh<3Vbvh&YpwOlpޗh0>Z#0>pa4>˱iCk|@ 1{'p-I+DQ2tAc _c4t)w#͵;dK;lz0N[~rJޒX\~?wiyEMZM՞]DI6s#j1 q{PΏ?AcR?fWw朞}Gw  e:wz֯i\}-vA$7zyeDF:T6`>Rp!I̝}&:&Joi9~3 5bڔhkNqK6b<2o)y#`72okt7u Nn#H!KHL+l"&7Mc#\0Į2Et>}ʫh3xNB$]1 P 10i^xpe<8Ed}XEWye6Grd?)9!]ۇM6hAkPdwl-{IB^$t{ 6k#|-(A'`%'ew @Ѯ;V$Mo{P;)&b^-0<5vrb:zh 1;^j0̤bM,`f^,~:yTC_h5^uspM/˺ ߞjO. x欔W\:t%oEjk)ŜZmm73'^mMp%7/ga;OjdDѡ>Q6i/,Tj%jNrUi C]E:4LZ[ptvC54z`m#]aWi]xq9ILR]ڗ9fQ# .Q|l1rshBu&`'aM&Th¸ݶZ7huQί_4ͥ6Ԥ&NmIpNHS7}5sMd_Z`J jbWiKZgja(nAѵCvUֈ/}I빵HZ,FJL;T?y~aĝw)yc[/z-wU u ޴ hsW-\\x;I^-:(͒%9#)bQAX dnld6DPy#'hP LIVMNk%'AMp-iix HI̯u|!AZm'_?ls6pc7A7\A`L:8*CƣX .C(ce Ɏ~_5FNaZ`r |+ 4lm!ƣ~w⅜\*R|9t=3duAk'37:a94S}E1!eN[Ce7*ДiXN9,2U:]}kY^cmd'cp.vw\v +յ*|{tAf&hQ[.)>?׽@8&DN0$ݗ~b :LJSSw6ҟ '9fh`t 耩#uil4N`pq 8k:T: :=F JKῖfhe ACHUn +q5[cQ#Ǣ` ǥ?}+p#D &;)Z3-CnA+a Ydj7Xj3C_G5ҍciܰW!T.-04?iLb-64qK Z'G:d`Pi))~vzb~6O׏ c߸~8;Bݣ=yt|v8BTN3#f\܀Ѽet՟ F05ShF9DP@μ$cU}.J{X(KC!GW]x:__]{L_Ooſg}ׇ]7^^J,_-%Y3 V9q|߿}RQ?~/] &Ӄ>ZAg}9*?^.~~z/]߷-^ƃg+qP}shnLͽf"dK :$RC{R鮪Y۽+RݫFg5X.&?,a?^eC}]?ɛerBf9ounhK Fx?x|yň;c;i[71w'[]㾷\pZ{+e2h'LuWKqW]}c?)[}8N-HF0E ʇs{o!5p?a;gb|rd^bo{FvwwswMʟޞŏo>!{]Bd^ے2yœ\GZyGȽ$|dw-/g$j}C|E޿ZD9>M+(G/:9ʶdXp^Ik :(ĕx/QZ k6K\0">J|bmB e|I$E0q*%?qN\ A*v cJdra \81PCV^7+)b̞ -9˙ws l0K2ɨ53FrpY%x'SA Z2IjUx)(:]*E+%5C@ 17PH; m /K#Ǒ-Y·Nt)dãG~XwLeU28Uk`R炄Ada`,c1 )3pv>n-/uPH[ c҃ڀPNH2Ӓ 郕ʵBS<,VYܣځޓK`}YR:D2q;V*nx!A#YSQ?`yt"{U;Ͳv6#L1*e9* 07cafEaݛw~2fum LV}eqm@܃0 #]rZ7KypTFow!un ke;5)"$$ev!?-mhVXCd # :"$Vy}ИQ8;v5àkF΃tH)2#Sk0YYcv؂bH IAqYY(mM!?GxC쐻#1Ȉ"!I"as-X`H[ 9YXBDIٶKQ;t~@,/+Ytg.1M`Y4 3޼Hm`JUJHc^yխNM +< L;?N5%3 Zb% @][tqMwZ`8D[ J =\;jYuZgtkfd C5Fi\k 'k7(aנ\:(Kf= s!??mSp|c,&RO`#,Pڔ wR [`~ EfV;AlHT\| b[faf%I-!d$-l MR,yhK1`@4lHu.Y~5<]圶Tr"8Ø|`ʣ U6Btӯ{qt o:r( ~44-&1 *=.n9+[*6\Y}sͽ}De\O!rh 1m%㲣Sn?LW\v->ϾY|9xenv|+}X,.oOO' [\˶JrtooyvBO)}5{:y/ku`u g=juc:EPEC:C:C:C:C:C:C:C:C:C:C:rAku`WeE^ Ǧ @貔D:C:C:C:C:C:C:C:C:C:C:C:/1&PǠ3*uKY ^xPXaFP G:C:C:C:C:C:C:C:C:C:C:C:/Թ]ǹm/7?ON~{^~QsNxh,kأr{^cQ,/}yb)M/F>\ѷiU2ĺ2kY^+o)֟0EߊuV%Yzgzgݪ@h~uq0J ^9[JXI!\IZ+=^!셈U&^X[%bpUXd2 !rXHKal%bG*[BVӳH4lk%ja!t%bRbgwRbڨA-z`Xuqk7æ5;Ոfv ۋ5E-;t%b޻J"wlXbK1D,ћJ feHVW"V1Hyu%b gJcD\Z܋a{Vn)V :^,r .E*WXA`u%b5/PV%b 7JZnuJzvVm)Vye%b؁b j+oe-bB(*kR0SX+N(ek^h3*R>) JV"8ӦB8!k+:D^l9Acq{Z2jY#k%;뤰\T"Kiv fKI˥bC^jWX%mE%bi*5 Q˘+^5(k[X8['L+L^,j+TʼTEDz =܍s\L{Q]|t9*Ƿ5?eAw0/~tq)"ӜMO 8ȳiErÞܵJ!gԕ3Fi;>Zev{N'j^?ے"-+[<Ibf\JJ͜e*[jn8^XSB2V"aZz%K;-{4ލxM#LW6tɶQMaf;\;-%;T~6זzT~o?SmR>vUY~Nm巏5T~oSm*M巩6ߦT~oSm*M巩6ߦT~oSm*M巩6ߦT~oSm*M巩6ߦT~oSm*M巩6~0c_$!]߂尸}׳f\y2R$z*ٟ q~ʭUߎk{^>-7*绰%:j⡐&=ydu j'̭&Ᶎڲ.'pl숇"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"xx(%[RC$ˋ"̾U>;s o;x]VLTQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]DuETQ]Du >`<)nK{ #+9Kq+-{gCEC,>Q'+`M/FԴhxOU"VH'D )W%E:M- QVN !k{9X+D,RjgKX!_,B1iz"=e.y>^^Lj'?y\iq!i!;gOP\^ 1d D*DVV3 2r+kAX\JzU/*o):x%bJ /Ub%z-]=~ROӭR]T7g9ep}߾9<6ݝO'?O0ɍU>YcX|}sPYү:y=8Iк_adNkBZ0n0V7o>]vp6q~eµhQ\r={}q%Y9K'6]ߨSw>&K>ۻyWqo>XF0Iso80)G1L๖#O`ٗy}o8Gh:'̇}x>зo}.ØQ{/8'ㇻm 30Ƕm:E$K &js"K=a۲Wcg&S<<;<3n'W95RʶQƇ5zQRJƣ56/{?>ݧ=Fgj= WJ!9!y}5*˪h2W]"M[^R3|:z`tmIv$h`kPȑzi9>Z˜{G{!}}FILk=,X3OMTLe*" <ҸG\Cvp\.[+'qK=Q7>V#Vۈը-cLEgm_TYeB4Etq8'M.U1ӥ"S}D_<|>SޤiDF]fsxtyᮏ4rn֖j^nzք;.꒎l K]jib<ƋaZ{IVѢܬR ;KN5s=W֟}4~n~}fU ='4@2˾[^O1lJ.mE{7`mٻY]͢։w\#h?{6$ `[~9&mNvB?-)Pe`U1H%>L9c5S]]U]U]];TrrD'W៛Yt[`t35rAO~PN}׸ǵ-zwfip^c6gVi$LO9?w~UUisqz`XCd•RFnԉ$MN`^'}j9cS+TeFg9c:9g2mYʚ{MHsӐLf侧njkt#%U*=lʼn5FՍSI+4llC\ Y'1ΗU]VBvY5%yBj8-r2W@^B*N>ra++fOP[?VS_TJ+8vö&5m/=lzDn~'^D_Q;=/kaYqѨr;}3rxuzz*]P6cl:B2DA h޼6*FeZ1wAl"&$b ,Θ ,x7!R[̧LhE^AUcS}&y#a\.gaZqwoog<|߾\LtxtXKd9sNW4 Ń~iGUkѺ[s+u+"Īc(evfNLDv~&ጷ~ϡOhc"Q'o3Hh!Kat@Oeo4}p:]nj@ͷ)G݇,>4U u}%jo+&?)UVN\z40k[Q:{W<ȝCr4$6+t^W0 SVöG6@k֨iZ&zG8ѯ #W9ffm9T1ҜOt7A6<43?bPyAȢ?v薭[s4ܶ:y@.N\s+FTyȥDa`m7gLW?V;-іdל#`)šnEM.1?q4(B\k TE&6'P"dԘd4&w3jiic< ܄RW8Μ6;v(C̕``uAB)toNZkdyVynBk;3x_^k57i^ҵ9g/CNyK߯+zѯ ѠCެ;mul]g;ߋۜkF6IFK D']T'l#3HrRMѝQwj\?gB+\\5[-8^4:1;$*8=rUxOݺo=1r0XqBj`Fi޳=(jt= {mN.l} I@ :`1JJPwl|6-.MϦg%j Ndٲ]pFf^NǂSX "i$/DK=12D:RXCmES.ӓuf|*ZXx#x+ 8l :%nU Hf &ƒڗ9tU]+5r󐚗ϼΛ4@v0 մ `%# 3(Y.0N *IȍL>&> X3;9_%!N AM͔pPe= )D%T{ {mWu!=hQY{{bx/nj&KAA3bTiʌG%Q,jA\1y~8㴎ӎi/Î,Z"\/,y.Nh;DW W?Cm[oE"1 >)3PƄ@ \MD͕ O'O9 D=@K/*[v|W(ǾZU^N^Y@MCu6|DV LrțѰCo@k>5 kg̓z]Ւ2_^&XMƎ/ PpNkp\uVR_}=[!_B*hHe@"w; ˙cFF"HH"< .{8ֿ-l!rE7;%XʇT@,qӈE Z5w-ҩl[:5wU8ڪ[d:;%y*"qΊ#GD*-}Q^\Ou,eS.d4"Z~1fIYR|O+,Fm."j9Eg8-Zd'}`8#Rd]vPm8odi*45\Lxտ£9Js2df,EOk։#!PP#RmRp^ )і!ED>ns<;~95=~Dd$ZPMC3iD.E6*E_(("˓r¼31"qrk2[XBhB5 hN {~Ur9ke՗sBŤLk9dLI#.BBX"hd A4[O[-k/v)=s>ͥ{y+4xrTfQ Y<ix3[NBYň(&&WT+xrX@ԘԂ8-} [|_Ԛ#jvj:F.3S*`ג FsDn_\|o4/ \^UsU}(g|]N {mΪV|/@sfo7+0m9}q8\\fib6$}^/\>+19qp݃R0 L8`\+(1'$tZg_*D+z"&:TVf*vT%cY0UQmLU23GÚNF4!_yJحuȈybRB ` a?CDž[5ߓ ;Ufʅ(^cCBDL:{f}, `ry:4}(/qv$t.͏'I-}`dF8W=p&咽%Rθ`-[ROjQ w`.VQUdXGG;Qq}5o%:.4텙Um{l/`\]Y.KM*&"d&wH\):/0; >pܗPщ}eL+y:WL Hƀdj6fT;?@%Y^qjF# bsШu= K?K*kkuU:+#Qz,QHY1DI[CwJd!-0 p$kNLdr)K0kѡDwI>ܼ}()ϧ|au;Ʋ*\_CSLM0kYd$>j<}KMvQl P]b{FHzgwʭ lFb6frC RҐ`| Aƭ|8pY;3?m:h窂>ey 蝳<0X窚Ε#N*{Wl{Z#$h՛Mt~K6fqwafT. 8f^O#pi۟g܀o$SN瀠%ZA%WKh5I$jL͵ӘDI=͖/z yt>O1_s6Q)vy{zkV>)K5߃IiCMc:]٨nZiZ+m+yzքRf"o (o4Wڳgiz7,9د]Y^y/WNe)T}CI=ZnZ .\I]Iss& K1Wr_ݩQo%,B_A1cz-e|5W~yX6E$t֋=œ-M1XzfSltGЄ)VZNH=!xQHͺbFtӘ8jhv?(ڀѤ4ۛXR h6ʢ튠xvw>hSfRں K -S䃴L#gZSgY{61]b;~ITMvw^og0kgv'CsHՙEӌO׌yLR+#s[ v}D]3" hef:v{MVA^PXW$ZM7}*D\&̂/Fxwr/mvj;U/|JU\Y¡X-]̏?tW~`;c-/_/!,x-b>.XJHvP [7c7RI%2)cŊ =` !ާIǫѦB²ٲsՎW5~o\#7?|(V˰޿[)l~>䐔/Uj/6墪$@#"f/o- :y }Xf}7AɼMÈt`RAj#kc$&l7Yԣ2ҹ|NpXAŬ *1VPLp㩹=x~3x8Mw_^UpDG21 ̹kyQWEhVE#|xݰ@L¥(xƅK9MҿO.gvKzTa!kf\&2sg<K~$;#36%mB4c߶99®=fh[ | w-N_VBo+mEJKV%TOiLbgUSUwRK*3g_4Njp՜1'=8ͩLIdA VDa(-tn"4~~(,#[҈"HASĆ̃.8rulSި^Xq| /0w5?9Ou_;6]iky^`oEU`6ac1j)?Әl<#88 ّ֮.& &hJrDrUZ.,v=|%!If⍷\ 4,VqZütR2Q;!M'>/+, q̆a\ _ V\BE iFoQC9.ݺe^{(b^vШs89(,ȷ'znWj$'r_j=\L[M(e`fSf(8*xAi7'0b\F)2KzTЧxsJ<SiLSt6(#e.K0al>,/a^8|( UĠ6A&{gğ!(Id%23`'4N,h>J<#;Ga)'VT(5eN|%A@ nz(aLrPݻGKpB:c>sh4[eܔ‘ggYzEVJ;Wv'ncP4m*KO0VW'bO9aA~\@O[K;??7>ǫ2ԩAcdžҹ۱lT^hw{Z r;euЫ? ^edڌ˜"?NdHzɬl?"(H1/mH"=&ZWD/0`i=&]08x~]8(΄B<1 0 4@]t8t'$ Ƴ w1R/m: ҉4r=|13 ƌ+|!KbǤ>J񞮠&48b7疸3"]s4g:'WfZ |(M6H eړd5e6ΡM 0vUӘbkUm#Sǟ?(:ҝfe 㖷{t͖[3h4U3l ɄQ,"SBтqBZsURv7Čn(y(_2J2+ gAOJrab5GQ*Ψ$eA .2GgEЪՕێtQCN!gwQ5́eZw6:tqQ28g\$jtsr[hi)%$ ^Z,`+Nq_i- 8- PY.V\?\C1(M2"g`u=f=PXB(Ƞ;h̸0#j+߃ݘ*`-g\Fqn ojj |cݜ}8 _U(nY ~~ ~$%s#Was/N?7aǥ?OY em[&KG_~01. aևm_`z~56f޻%o׀Э?r@‹`0WG:?Vnn}aZ[?Okߟs6Gڅ ﷏k3X \ԜX|Zmq;[}+8殨t5۬[ҟ6z-5+_wi_~8 ͍x18>dzCas ćRxmEUɸU܁̀WUx]Dm=|z:QqLY؊}&+^+#UOxL?ޝ?9XK?tK5H)Hކ}/Bc^U snorՅi̓۾>swviXw/?~|YxkcoML9噷)6Ne8 rM3u XG,0kT-_|{*ڶiS[­V ŧm6cÍsº@ z l)G,'_~{- g[{kFw g/pZNnpJ80c|y(M/θ#j8|G?8+3(l\JP@5j"Û& | s53+ :(fmNbkyIʈRt)`RRqCٜ\a_-1. c2UC9rYD[wޡq}RkUycܶ63Kj OعG2=HʲNdĚL0,>'`GYYY3Y>Bͩ( 4.y[`fV K`,b5QCY9g}&%/4)$c %z/.2 s3T XSKO19iL&At{u8$ VSN]KɹdOٓ $Wu&HFUB1N2_%ɗk)v{PEϝ#YS0G|IdPJ*d\amwf9Yr/Iq>Cӧ=aL6b'yq4Q)m\|* 8" .hͥ w1>YFT?{ӣ=X9j2`yg^e$$L2(|H"p4(,$^^t=IjNw_(ULT&"&Fo͋:Vk0l$ۇ__oW:& ݒj[m6E޻ǰX9C\}+YxA} N:[ν<7qKk6y*64$p;fTJkPr \9ur K/JP~0A.[oVH Ga㵟 J)3 #b0aʯ萶4meJ2#qOzUmͲU)z<aUYލx H8sT':@PzEA0SS.;rӥL\ŋQ+Gp  b1nZ(5b!=d0L]e3V߶IWM)NsǞGl֜1J`X#VFg=бK&R/+Es- V:!$%)wQ+b=SuQs4 r8KIJk\wr/M _2ң ٰJb i-I"ܑh/b`}_:,ZZTN[̀c[[ʃ؂!X욉p僧IXFjUUK(g816 tp/f l  n ,9ၦ} #I-bP]E`9Lt#{{-yzň_H K'lIT?PVW&E3qfW .5MԹzW)ݯg__;>+R Ȫ-}Ǹ:}mJo[ʻ\-nV}(0b8F֡ od2Ɉ,?I[nJ2ݍb]I0cjCWo5cy0'PtPp87luHj.rb~j17iyV6"+g )1H^cVF U`\7>rݧQ Q/B/CcGjz,Z̭sa?E4G4l?qn"<\ 0%?:V/撐W,^dl8o\ù"X;{1!YͫƮgg'SC#iGPu7_( h@ƧƈxƓYWKMs_gcU6o< `Hr 80ɳ^?kf{>VzQiH=W.1|@ &u0Ƽ*6@$$qx ޣ0[_I tfvTrf77/C\GԙwQHduU/rX&JfN7xk}y_v~4XmujM?oU/?ϋ/ǰ@{;'G=v)c׍Yy@TzF|pC^f`Vۗf(/CS:ꈂualpa"®#{ctxĮz #DF8nV>Ѯ_Jh|a]Mv %DJ}8d3"θ`)qu=.蒪yΈ?Kw^6[0!v`z c Y'V6[D!v6ØQU"dj[g\dEуYr57Zrآ۞3!屎:PnQQ--Z qI!"ޚeZqsߵἀ1y!CP|U4+TmoʅC-q:>;H HP%7Jڞo !>g}N|*b/r>c>knaLHcDm;1gfI e;ʈ ~W?m[e++0&QV#La'0.H̫>bEהL_6m=@>B E$21i_0A8Ž*jeqdI8]ݨej}-:ZS:.0&A4Xq#á$E9CLKr\!\cbd6EӭLa{5H_M ă^Lc\27)7uB0r{m`}Gφ5/! K0? xSHMT wi6} !B[_0qiڽ>l a MpR_L,X`,J*U(tu 0s ¸w:Vp?JEF\+ØzλESowM 0PY&çF_T\rE#f@3敇TIGavu7džԻ:bplx=ITɈ F?iө@5j^^ Va]FY{u'j ,u|3*ʽ?A|4pAy L3ˍW]~4 ozlH'_U5s~c/'VT3 =ы aUdd5@;YDݏ[ PG?1:Ҙ@~QA<yHc@:~f90ó8" qàr [C؊rPG/?)OE%ewE6% (RS  uj=ow8ȅ&}R;ъM|LrHێ:Zlk^0&9&159'Wʂm4F+sVu|bF[ӎ;6 ȶy/f#Ph&DԠ<1V?W=?UXZLo`zۚiAF1,bk Е;_rpi%uZ@ns3 .f _ͯ;jw ՚> ,kF!4E9L ^۷}>\$qWPï a3ӈoY%Uu@A,,Jxg\v\hZ(ޥBIC$z!xL J77tjnjgˑp+*ˬ̕g;v !|ڟA2e{* ;XwDhxZk5Ø<&2vj+9S;>?/xYv1M"#f]=Ʈ~/*u.VuΫg}NՌ2ƹY̼&]՗ z]1ہ" )i!t6 [ZJt$葈J}(yM#ѝ(nzրi,7hl<\g@gO!Qu ׇ.NQRv=OK2vªW~7V ^pgpv`V#a&s InɥW˸թy'r[5ْ9ǸF^v "WZQ7pauz*V1Xu:TԍSiߎݪp:qR~3/XWeW9_כae,67.YUȠV0wa;I]ȞM'n(Sc<-bVYiS[;ڇ{kyaۋ[lnzX_= q.@MeAU=_z?蜭~qٻFr#wc,l|d|"meI# $KVe%YbU9$,Oz5,p |ik \8luZ Xav0i6 FB+L8boaz/0~zt-`G|>Ln_soveT8pne(UŤ , CK@8JT IJipQCOm3mx<4.Qe?p8ǘ*p=ԓ̩ b`&!v˨-ĭ+)^X,n02DŔ9v4ZMZxp/֫EULrܣ|ecUvPMGpQG)y V& ii[hXy SuV@N/N^VpI+_MO#?Ns `idxM=G{EBu<*_0_e xJJFuW;50(_>op #[ |4EtekMphܹRoTR͊5( ҲspF_chE+'U CH֙4 c_9漚0ra P4N3Vy@,8'> cQ@O5$cp5 -e0*z74=٥>[ϱhv #%Khݲ}٪Yۥ[UszืXOGǝMޢn~.ՈȺN"`r!q Haxt&/m݄cD *@88XMW5( Dm}l^YۥUO2ZML'-"C`BJ|DuI88c͇79W [}f{ݍnq ;&O gF X^&FYp]he]Է)Xd<>~Q\rq0l6M\c a)5kJDD1 0[l5.Dw7"\r@adZ(6V!^U&`5!"xb0NHGݴmy_sr,# S)2k 9:c xUR,ReF-g5( )?tXO7BYX@X1Jyx,FYp8!X$Y)K}.~_avm4*X$HܔY]u$mM,8퓬 Ѿr0T+ŒYy={B 3bUA2{<,Qc.!M2"  qN2`i%-WH Pܢe1cQ&܉&$h,B1,*(P*!ҚFapfH:E<lld`qU*9Ƒ܁FV2^N*3TJkM`-w`R4wD(Ʀu' @  XcPX4f-곎hѝ3,*]}$î We 60-"Yiв:Xs{?M5?_[c1b9 pM{mmE' 0..33"K82~]%2*ʄ4ã50HGF84ٰeQ 37-rh\ج^t c Qwjkf_!u\,w2&{Xg]/75ѲIKjGIE+\$E3Ц_& QD@B RX &4ڻn رU=^Wo%ma\ۍ"YM\Ps, Gv˪gkh|ܯA`sޤ/U}"-v'pnJ`PU$H4]v)biR&j{s'<wóx%6o5٦Ԥ׭Rw,(Z2.LKZ"NzZN&AI#h8{\6jК>?Ttf`ӿ?.?Y ?', l;ө`xE)m-ʂ~my;ҿc(ރ ,Y!ͫ|<M_ |^v<cЎ_bhzd9`ӗ[ />>LњO7>B#;M\axa~3;/>tLU?˯ԩ-SMݺ =˂Iܼ ױӻ5xAU>ֳwW3(<ШA?|u|xgjn<[~Wuzvv!ԧv3b+&}\ (6kjQšNs$exF+ˉU XjOyyƅ[uy=t?~yHP S$߬#ˇڎ\~Y*_z2Sl`A`hժ}0,3hڼk~84֟H< q580|4k?jd!x `j?<o}7Y@<3VfdY hJ ŭe C@jr`tG-4bIQv OB#ΩG#n_qwcz3fak8?N3;ƙM6^bog<7^Ŷ͵';6$7?MH((lvGJ% 65 Edw84VΊ>'"1$FCg pǿoQYmc唜9S.rR}Lz\Y^T (͈5gy!c(%W wGP 8\]qj ۠=CgQ4QQ FV540"c`DٗgP:z# A4IϊqjCKG1@taGd dhb"s cV HL%?]85/8LIg)=g01`*OƩFk;|tA0|(AQFN;px&K twPYBtgΒrbC g:)r;px&N8| =RgӭBaA1KΔq֌m` )D Ѐ^c:+wL;$Wpރ ty1FMK#;y,h'igu3@_t1 L5%P#r_4N|%ťx;K$o^HKFFgv١Ɣ-< s , 8cgrܗ @cG}9ȇm*Eq(ف3q8.|pJSW@W^`tsZܺ{gӑΕKTEЋ9~8L ]I=bNW;p%p.qzTQrr;px&]>dbAֻĄNڣ!lSO-)6{ >])/r=g( gPq cL:8Ts\P}e8LRDmݦtUIɺۄz<_lA /zSeXg{w]n<Y3 ΪKJvs֠w.k9 C<+{`T tǨ H~ Sobˋ _E%ze^j@j9KJ9}FC7*@bph @q1hCЛ A b@/3+k[Ǣ7|o<֋hQi@X7FQ5ciùZ쥩12'0n ]D<ƶYшN6 ;RrrCwq  XryWg}'G|2XӦ۸ꟾ}i[>in283ҜόZ}cDT -t6̻N+qPm佩F)eQ,滶cֳ-\勪anu=Y0}„ݼQyIT(%QշeW zgLxj]LgČ;:otӭOˍ8XV&e޿ !Y},ê`Me[ 6p%r*ɶv]w|lp}YeGcYEƤ%1<՘YwN c$8Brn `Nb8v< GX e8!K<|nἣXObfS6xeGOj¶? ܷ?8BȜC b\.V١lwB;s‡룜cHSb^j*sݕNKc pa[yA?PcFϟ΋ɼ]d=_IbZm @D$Haq}a? H?fٿ]6~GٰC}?XO ,?znXw v}pp:zV%~`B .GvM2Qiv?YI&?u._IєU!R:+mMX\6a(U<3snj M L%"N3"xB$RPjq uld:I I9$0z:wvF3:7Yqj.FMij迲|=x )$x2[(h|'ggN  sg~znke}fSmszsX+Xwy|rl6rgpMgixAk}iFS/Lܾkܸj Oho{z [|VWU:}o~m?j*X|ݽc{ISjn{IR1&hsn^5wwxXCFrV' e]lf&~|n77oycaoL q7nf}-ӻ[M`w /{6[$S;+scUX=c!!5b l:9 04:ΌhkQZ.pkipA},h0bS?owQrn-A90[N7D{ǻz`8dPBnS&&Su{芅6,ti[}awxWd(C@$HQ[Et@InFTDp7D.&-d&Y3-^_!P)J4Eq\ r|qbtRJ<l%dؽʖ9Es=)KT,,#jT?*L^6YCcIF؁$8`RQ} ¿[-ˈuaF |/~H$4\q"yF!QBA DKi,1ё &IbއmEóm_ǚ> F_aH6BZv7-21.*5A,JFkr@Do Qo@7w:Io(H~ȩe8ab8#cJ ePkHRd.86UI\_ E0B'1L* $$3[ JKScs*2 ifB  $Ě+2#bB! c;&Lj%O.W|fn{rE+vʖ2VT3ۥ5VrhTBC9dXez2t4IG6H>?|^0( 58 #{^ yc# #A('kKOu^r2"W sdzJh\ݞYB1PlCW[SǒvݽvVzlRԡit|fILACf}dW-wIZn'~tmr[S#-X6M=ϩJ LjV)G|18t3l=ߛV `>"־g=OIͧ m/ﺚ`fvKegDv\"A^;ÚJNbj}*)W፧t )+*8ށ'F?>O/=\Nʖ{aWEߩر=흊LŊn"va%NM\$#H=$?Juo64Ժye>R"$A8k5&LJQrMտ꿅խ%_Y) R N)u{u=* :/s][;4Д$ 8Ēǂ(ɥHF W'LP!HT X:/b%1^̫t㹫{$>SBL5h8t?ɩwOzvm0ě+>$0$RB:Hl-`HPBs(b?( ¤{a]hB|XӔ7RɘĐ(9BZ (TS8]̞ϳI{ic{3ɿ Kv =lǺȒ#3 ߯zm(uS<ʇ$eYw+4\OB8  F4Dؓ6e^PYLJv>K'[.*hf&bxv!(F&Jy`8dcky4T5$,o&(qHF 5xX ENr2vmJ$ćiմN.H#^W [/$6680$SyKEӯ!)ќnGS4W/\`ܦrI6, yb qva3I P:i}n9w4\ԱZJ' и \Ϝ_Β?{.ʀsUL7 f!Ԉ\xG0 0->wŪY89u1r1QWk0qG %tVK4I8;;QOi@tv`iܙqkb\l`=W38\?7rNjwHK"ۿ_̧\F@ob4i$ xĀaR)~z}旊`A %~?Wŷqu*AH v9\ޚ[NY|^i9VVrI268,yDGghM >͉'l+JBv}aӾBb_ZOH$T1[ zۻ+A8p1 F`U1GtdUuzM;BvJ S7%4+-Ha%Lƺ xk$C"!e:@\x_"[  R8xxINz%aZZbR%OLz0S'+\ zf(!JAA$ǖ8. MlPT8&*X73t2sOۀ 6BWq#gJX0tpC9P#Er-R+- t]5/BLkp)K*00Y !%PC׻^9BpHhUNო#6\)03%FvT:d)`rHpJ7ɌC&a5>rxaߗ)'ɥ ⮤/^#ʪ,YAeFk*+,a0+c4ދxu(U Mȡs +8h̸ٙݍ>?lORͅ3Ah{P\׃Ȍ/gYf~=ˊŎn&0G"4nq$HdЁ^_WQt9FQDLQќ E~I~.b2n3 U`#ndT0,Ct&#-TtߤG 02#ףaZy87UJ‚-%0YKMѲi7rC{ FðHAǚ3#EONF!U ٭Qme J6Ȍ'u~4O|p 7YS̡-@rcmkEkODrE"Oɘv1a+uܭ2JNЇ()ǘF242FQ@%`j6(E\ ҥ kl%;ĸc^g59F NQVDZXxI%8BiXv w!SFNRc?g ZFހU1abD1|ʘ.fFFaNI:ۣe UL5H)#3r:(yû=PBbX?V^-#3r$}Rd숥Lnbˈ YY=3B0?TV1Rr*&WM Z\cU1!GWźm2k=߿Sybt}ޏH߳>[`4-&9ad%88t M&=Ux%%%OOJ]m4egxw~ξh &EM DD+?aE_Dw1ʛw_ <;V|m }ӟ~m_ɏn,}7oI\6fOa\Lh F( %p|jћ.lO>L|0_O]ٹWߏpRh KbI]Lq1C{yf-`u]}T*-~=E-~[̞O9QwL/]w/j}9,ϣr`}{8\]?[J;p0=4^kS9*3ͳWZ ܗw$ns)iU!sC.B.3syTZۘjŒ (skN+9N0fKp(HS)bĕ`αrxd=| c? \Sͤ@X{;~8Zt J؃+4"bbލ~V!_f^Pa=Tae׋!K$㥪)U1Jhf:cR$cG4>]3ॏC1" )s?*ӘukFL@4*bր`#촼~6p lzr9˰F_ +L=5q]YG{;Yb|r!gZ cTr]F"O,NWǘ%^xGum UXQm]?q8my֡H 0Mhԟ}4͈hd%rfM}՟~Z)8/ŹQ$)r-CQC%8 4ǃe=ǨI F&j.y5%/NQǒP i$$T2$2j[@- [Jyf>&ܶ@pب&o\4b}5MY嬲XK Nme78#qT`ʸ2GjţmNBԱŊ^Ŋ^bI,V9*(Q4(ƍT@K18E 5ԫzAs V(Lm rً;_`6KvB*n؜튣.[[ckqTKWFʤj 2AZ+F㻛cpbuysu{[Ic\C@K+|No\\}A]A_E*e967w}xp#b -GNp$ ցM;Bh@Y;ΩK$†{9jd,&҂눝zIx ?_Ƒ`y *7آ.9M/_C<ykaNN,#(^!^bT .IQs'eIluܙDDZ8Ϋ >S-r\A]CowW;[R|W鋺PBU }Z< x%H`.P(Qc۵`(+#0V7MPBr9/㴋Aո[/!H~PxHwW2ILc Ԇ[X_! V+M~S7'Ёք-}rKumv4$w֨Fe9^#-_^X:hiRTl|Нpi߄n.0iQ]9lZ^ۛV]o%&M+tzߍ[}X}bGhm&\kљvRVǨEQչ(>xw42JXb8zU?[ܾZKIG`L~9"L%|շK4UN~siUI!,XH{?z)7Y{y|0^ K*v]|K)_Rҗ#+l|0^b? \MGhøڹѸ5FbqI,3̊Qz٢6n:/1*b;'8j8,ܝT猪ؽB5ڰttx'/ aL]l8A삖[w$%ժ+-*@ iIE Iu+tqxyׇO:)+trxLׇOb +tqx-0]{Vy;칮0?BV)0?B!jo+T nވ|bbH4~Pr㼖!UYQf)T'b*P&qPֱk(%W+]%/j &.'PѲ;{J9J[Qx-S20?B歴m^uJTIhb"D+#2 =wJ6F[RfE*e' rmSc ];Q裠xJV0SfѩGl~ ̓"g).FZP =lGB =Jc@$ҔyPæ斳Kc̏S|Ffi.Vd>h+c̏S2])I&Q[(WIѥX z08S(`޸]6{ 9% D4.D蒜k-$&S(-+?4ܠ>x~kyYG\q\߄G\ =\w~fSG)jȯQSv\2)=MЃyr;݃^Z5 17Z\@@GM)#)z0zWgU(r*CrźhM ];Vi4oa(Dcv-1?B}dQ.nJ\CIw#d ϥ gުn3Y]&-!Tt> )AZg:'G0µ8Vj]_9%-)Tbt(|va0U@։m#P Avg>hdZT&D8mtڍ1?NZZ(qG#x7~ TF B&o 7ɪhtKRX$f z0eS(AFU%M0S(ta(G2۬xRZ ռ@hps(t`(d APpe&!DU S(`^ݦiP)pkJ6X4cFB|Zws!T5QJ`2ܶ|3X/ =z'9)[c"#@U򢥔OK)z0EiDA PRMT>OLIڂ)z0]M h%gJ*J+ִ #n + wR"--R B,#I =3_D` f8#RqKk(yP¼ս!mAI^"D2[Oڂ&l*Gl~ +=ëibPR4!+)ua~kӽɺ>ֵf㐢T/V@ͶBF6 :K¸DtIMjck Pfն7vP], l"7UQPI^_Ah=5|}}kٯq#x}cyL|{o&l﶑ iw->;ZĒikWUe{O_m ('^[AN sNtO fusEOp'>mՁЁt+)عO]pT6ڏ](t_]iՁ]?ёOnj>vwţY~]ܕ;Wd3CmBRys_?NjCzl>[MйCؔJ{a^601-P5}TI:C6B6*#jZ:E"kjS.cP#xF ߿,[!lz܇e ,sAzJՅٚ?y1M_=ac__MWϾ5Ez.٧#7t+KQ[Ʃڊrs%I,R$RC9,]h;>Z)y, tv?. d|W'SVc/፞5VAMi2c o)OU))7Thmnt5PBɗ#ejg(SU*ЎO'W[tTFǬ\IXm)]y cd 2vМml[!%BY.VLhJ(y@kiw+WAeB=jf.vR5mRohƑA GPy|u:ȋM]pu1c6 6 `Ѕ|Չ3Ep%!% "cc3jA0.娹 dK LK n OLpQ*JSQVGRrJ}\j* Qљn (yDtW>jO9 ³9VS zN"whk.!O6J )2lGg*VX)x }+O^ӺD =&%A\yONFZ^D׻4`%'Ӯ (24((vi#2C‡qe設]'^8Ҍ=oCrhN%A`cvd%Lujh+(N]>/F-Ţy7mW(i,lAr *RS$Ps9U 8{# ψGb-{Bm6"erMλQ(#:k:K$Ps$H¾jY"S|ٛtPoY%HeHeccT9aoTU>U4YgR~ ±BPe sJP9ISAwjUFy LjbG%֨Ş(@mIJSѢ-!NmR.Kn>BJ`)Ƀk,2FJda)MZfzfв\2&Zc3LIBTon(9y΢{sK!lQ(7-.zv{Zs~|z_ W.geL)8}}şʷmUߞ4]<ݾ.~}Eޑt;S9ٯV'ק޼< 7G/wddҠvlJ2$[\[n95%yZ\UQEos)hΐҺTP)ك ~Ec!W4AB떟a0} 2|55sRvSoo./Nguʟx"N(nI8}'wMB<oljt7s狫s@oG[m+6bh|C>i]_!̆ͅzě9~Ws^lKd`$/A7;u~ay8_.W7o_7;.xxr~݈:N㣨r}xt7Hwa7uQg.G ?l{}= SNcx/9Yd˲7$2s3UMgTzCu`V(ooCbڷ9S!%i :[/V/NOA|,OY\rń u8{p^z7,kcvLf\N^o,M^][W"iTO~dc lCV$ˣMQ-ČW9`Ij&!~ҟHc-N`F^[ %4oEVJhYPZS E3զ%'F%}NI]NyK*FdS**]ٟSKalO[49:°L"aRB6L.d&T |'U%$_ 5$ D=AjF:S rm'< w: &PtG_OqGOyPj11ɋק=y"=r@x?/?T@<嶵:T!dV1TPQV)GV*\(V *eX9/J7M]⻗{ sP9Q#=_?;-&AcC#}T2Dh\m ϤT+*Wn<Î d z3nr(p]r\[e2'wjP%v4!6YMy:X8{z r^Xr%i=\cs'\ˣWnG>LOq#_u.VJ-iP*{xƤ BdvG,r?Pź&ZvDۿػ6$Uu`\@OkTHʶr~K")KLYM:IzeDSxsۉ&/~ՠ0ˠ?=@+F&or` B8IJe{F2FHH5pFBbe@N \L)d3Gc]X9KZ,i=Ŭb-jUuT`lt_t%;.FUe,g9l},c!SݿCZy|SO`{dF KRx @9DŽRn&x˸M:2,"1rئ0B"V_"D4.ߝ>.7 a i"U'|j^O'a 0mOpnd̷q:>!ڏEug4kke-e` ~&BP5; QmQ'(INtmK೅Zu9v)(mls<cfB6Z_hC/L:qz:^40Z]+ngN&{vrܾxQi|1?x#h K{l7؟-" 4@uN)EǸ..Gqفyc`I0sH[]DS2l=2_͔N0A0ɵ .e8l O=cF@^qFXp4m w9{kuXPX_g ݶK;NXeKMl^4|Fb-HL{ֺBb[|t[f]"_?JE1M+[EĖPƽuH ti'So-e<ȏP!I[zفğ O9 D`q-Xd,`TN wSpV G Ir2xA ns@uM͸~A*%x8rersrs 1]IgYLKX7=}˟'vNk9RK-6wSGPozSGr w53 A-,-r≅b2qS K諾P;In#^*kܷ iAdGڶ=.tn𢘤Si/)t6}hqjVj|q+"*MHL LB>7PT\CdKm7ӆϵיu=T_d܈Բ@u Yg&u)beRKq`n7 Ԩz{NPiAY'"@'}=}LیZQ$8Ox;rt}r?}ܮ%h9E˝E*S0$΁mݘͳk"KݒW7@3B^,W7`WU7~'M}ׇG?Ol493Cqx?o6y2BX D4J4 eF|`șż2yu*<(VARklR6rww([t9 m?E-_`x7 ^uzT~M{U˼fjifuE;)`149ĬW逰 9HysQ b vSYAyn}0]Et]eց`vP^U/882S^(dt1Qs 3aJ~^ O1L AHRVa S띱Vc&ye4zl5BZ": >pt|{=:?ьөvZ?iʖ;t%,t-}& Y-]ךxGӲO.֌"7VUH&C Ȁ.&9 MuY%|ԭkqM -7C jhy=x~4ۙ~I<ηt\Ä\`a[X~t{[Ajd ৳0E N[lBX.*'Wrg[X(ˋEa8Mc,>V燑3aRz^ kSn$*ʾiWWDGE*Ox< ZMF&k8~ TH_̡T9L~-ͩnݫ'7#tRyWfzU?xWsb>a/z "`7dwoB-#ܵ#1~OmÐax0чܣ`*&M>nO7MNQ ýQuS1ﴄ}7qn)7CmR۷ݯ'\󽤺M:$'o>҃5?Us^s}5QȆ V{}B_xo}~_`.w߀.|/Rj CI1A폻iWCxT]]κ^%)q9V\u|(m~i\KzpdclUbQ=>Tv6o6O<mcGJQ&iu֏o@:MTql-' UrR=2B2 T[CXe#NX1zH`&׆k2hV )Ï~o0 &Ϭ-w!j0߇}[}#ĒcY,s\v; E1TWu\muʿrv xkʤS oO[^61NJz;^r-" `BK \ 309M?x!-'@syN ?c=fc,8k1̵\OZD~-fR|- 0r5\ #W0r5\ #j9=W02jFajFajFajFad.W0r5\ #W0rxx+FN0r5\ #W0r5  2L^.٩L J-Ѣǂ$-zGG=f:]X2(k+vt.`0>1еl3:i7՜N^}PpeC)u_NZQix`,R` e a)$$ǠUJt|]Y,%M)0R@Uy!stZШk*j+7s`rnHϡ%d4M=YE9-E}t X%m˧ڊflZcFGBGR0!\J+AY*51)^^.|P)u$E 4ke\ӂqfJD3*Z9&H{mί= 1l !{aI/U 'MyTrn.cigc+=lNʼnܐ{*-?o[Lt %>&b5М*j$jxKv,#JT@PF1R3!5fB#]QQ7_>TIBjŰ[)r]]]D帴aY>(gZ8B%شԏꗁ|KEݻ&o{~ZhR!ib)c$ERpWjMNU_zJgdwFO~YZ[9=5ҡ}61 *$ی1[ϣKPwbJ8$#¿Ӯ0ZOÆwONB[ σdـઘ),`EN\bAW5Qg&*iЦ wv[e_e Hq1e ` ɄpFA>@:!iZiRZ-v;ٖs-Tla)@}+prUMH9fPʛmԛ$&9hċе*+67zMGc$s6L+V89˼IRR2.tAX"ZK T.焃 5Ƅ̽kȼek솒h Jjq R;!UKc?|θ#1 q JhuN&jUr7wL=<1T wD15+Y;mv:f'Â6dvL H E<bAe\ZڜS`ViBAL#tQk)n_.SϴgR g̅VkFXhǣWwc{n6@E\qh8x.nHޤX * ݄ ^*:3 z}ۑO}Y'eW.w d:$`ԧ}pmdUA;Sj p[ǭZZx ᧤~b;-5)fSϺ zuud⅕>0#Y~0m馫WNYJw%m  @dKm6t}GdRʝ[ wnl w&TLҬF{#OCl`$r{ Er\Ȼxp9Hp;*~ͬo<0Iuoy -;,f[?TLfbbf8p_h/di*--օ6I&EmDQ.p8m)pBl kf(?<>9J)Md Z#WA fN3r[m5W@ J:ﷻgH`ͧ=k7i6&8N/x;Et\ޗ^wG*׆55ڨmh|r>jzͰZupM䁀WX"1"DsW|vќpɤK $>USS< AdA3IjjحY1޶ۭġgHܧu]$>q_DV&r}-!C4 */hi4(Ikѱv58X&7Y,c"EPn+:H1瓲p)ӢBȗ4&0ɨ`zfWy7OՃ5x6U!zڶan7Myn?3>tzzݢv be} F4M~FsSG~\Y:꺾igs~x3}nb rZlA׹ kٿ\Who=thk8{Kdriɇx>dN`c$1KoIԝ=JF:_ @KkR(F'R(", o#Ą:3E4`)SRXkȱC{w氰hl^^<⻾~^2-,Dz)88fr۫+|ZQq$8\!ΥxVSo;^Tеm \32W`#\!s1WZN\*4W3j\!L1W\`b N\*;t6͕>M2Vޛa/ ?a FѧaDF<⊎C?Fô1 ~\-b)|4%_b(ByFJcsz_m[:I,o P~@P ]orӿ頹 Za_VF h)Bx*b]̤t9MI ; /C"R*PLǷIJ0Y'ϐ&&e!9 exb8%6HPHƒ5Bn:k6ibmԒI4%` w2YN5ɊvL+,7c1{u~ FѰTs--Nӗ>MG* *w8gB8҉%tZ( .Hd/ /z߼wYy}AViKJԉݺNJkoJQZ m%I "5^alXX ]G> !1gf&/|4gIRnzRT&죻jb#WQy1?jϋ{8Wj;옿YiW{d/eɵb*gPw~(wVc}{7H6,gl)Oۚ iϢ/ Hgc@4-6D3 V*tYZ0/IlǮ9f1c\<%F#cJϏ%r,&<~4w,1Ȇww,6N Ow>h?zsT`\5g+m'zvbh4%]z{(l,/miiۺtϞzc 3K~ҰI&Yyj{]AR+Q N55a+n[0p7>]嶝[|, \T`VGN ax`2Ogߪ綾mCc0_`z[P+@B joq32W`+\!s1WZ~eR@gޠ%!!ogz_~G"VW fN@F)X-BƬT$OV(Mf4eSFiTRGBc/ԃ>,0\jE: nLyWE\)Tg1!_>Bu6]3~wL#IG7uxyo!KB.ȇm'\Jf\KQ&bT ̴螅rwޥB:Ф ldpQsJ,iO8#3v>qB.sAZaN=*r)zxC͕Uo}tFĥRODLU) Fj4@H K\X+``ʈEP\WH9!30m.=Mٞ,w@J[|QJQBJÄg}H[n.<1#Б=Y(Yhʁ(9 JWfK>&dp('y kmm`4}XJ.-떶zP IHa[h! !T#^#w3iԭY_E#k4gtZ>mܛh*ٻFrcW82>L&vq2%/E[,i%yg<@).d$۔%ٽzlufWUdIpRl[cJ@=oPѠ0@=hQTELD-:7pbT(6}?}m7 Xsv\c]yLvg-_oґۺNG֝x:R/uqt碢}g[ThafcS%{Dn? n}/cՏ nR}nfx~U6FAٴz. [?mr~crr(!j %ՉKY+ztM4j\}5t K~Nufl gejU2uA颫]+:8Y~wIIY wb@@83@ゥ)q'p8o2e 2bG9I6Lq[41 GьFPeZJR6Qռ%O<]g2FTvEd%Tr'z0CLLl6Y'Jfhm:9e~zk;ZK\77 /{?7gTF$pTD왡$F꘬gʕ U/.*3nG)G4㩾P Mn"M/;mov|7̓?rk> x1L(af.2½ѸA DFcU)@"Alm X^&*aaI*7 C@8h4$((?p>&I@ڹ`^YC QZ )899G98>  BOީ<6h-iigma0E涐MvQ;YY~фj-+ؘ3 50{dE0GJRU<&^hAמ%2]pP Xe`$$O]ms_9$%Kd3 āQHl8)&ZAyA(IQaS H]4¥s"zs& DUI !#zj1!15ϴ҄'V8RLP=t_It[Uޫ9O*j >D_?-<Ĵ|dh+/OG'$d@}e2+ϙ%Lo`elg R[TW UY268Ef!HKpkiS@j 1 DsmmBrc.OR@ %' V1qv#,+V7Pͣk{o^8Q nyi6f%Yk{J+Fk&FPi:p$:o㝡WBfOSol9ZeŶؕ}mootf4{n2,7/Woӟp)M /x{3`G7A0DI@.-L%ʬbN0pHDbL[loP-c>ɓsqti z,疟tt}}zIP2xLy&EH$Dp7G?o8* u]MMh|]U 'Ao)0FE:;: ߓ?! YiDgS2칔Ҋ/hmAXRF ?/ŧz0?QHAH[]kuY"ZKod>y 6*r֢tՁZZJ"B8K{E锨G#,Jrk)QΜf7m^ n>m"Â1jzc*͘:儹=fJ&IDt_vǓ\ȥY0rALWW5:FR=w4a%I JD2Ce)U*hjRy]@Ӆ^\^f_كf9}F;c-w$SHκJ\}z]=} 4o;2z>/\c fɗvIr=f@ Ӱ caj&#5DiۛkVm7Xk[c3n-se sY`WlXQQ/ 6@W'w*1tPRK- ޘBmKlӳ6q;OZ/wJ^nm tvi;A;gn[Cz{[gr{Dd,m6W9ܓ`8>3s4>/0@vJы𞸟ZY)㩚Fw==ݖúGNY4Yg!J}=}sZՈmp[hv'fՃD) $TPSbV[P1ZɅʞ=vy[6mU}XK6ݙ yfXoymءZ݂}չXyzݦ)hmp.\!C*>֘\y/Q,NZG ӕ,T+SY 2QHxܧ61= @0IK I=':"X D7M% ZB[Wح cm7o2r!Vvr5K]gիLGfL"[rhT_\a .&n ވ^IjA_;[jwkoѷSԄr hq$T/*r? GBqY YSb圥h)w6qe2z|dY}]W+᠋SF>h^8<ƶVZDIdưHG:l0a"C T@`)<IeD.4HY\i=2QNDQ>gO_R^XuqǃQ&P=LOSߞvu7UuH|,MU8&٘L17\SZ?GZqЈ:"Q!(Jg! s \z M C1RYCzJMw@ä;l9-_BݙG;ۋӥgdsy1\¤3ZB#@OcyMCO̸pe~^{JhooF&8J ȋSQB965$8p{w : jvmQpJ! wx6 qH7|׈_AŧT}^hu^{w wlpd5kDq:R){݅]h[F-W8=5Gϝgݙޞz+?>Nf|\gL{anܷn?Wӱ~lPN \1= iIQؓ{i5w#YVOhGQ3߻Mwh앑gݻqtFr篃QMߎN^G/hg0llT7ZnXxY?w~w|70R/ yc'ο1nN5ɻf 5G;;B#ï?ǿc×ן|̗Op. \IwTAT߼kn>]ےO==e#7׵~%5, 7v!1Ӏ@Ӻ9|U#+g2)u~n4;yqFui 1 ?' w2c ڲ?5.֏`gSzۤK(7&k̵lKHB$[ zq7-uؤTB !&a/òQKV NbeU0!L8y| Fߣ'EL59n $CX =h୍c!m *rt3Km`{&Y;¹KrSi )uHvy [D>;޵m$Ekˀq%&64=ܗ+}&Ȓ+I%)YEYC~w<$(H9wX# cC*ru gQGeɧV5\-jQJZutߺŶEZ;BZ@v|۹*OgF׃ްR8@7*$T{n jqsԫAEXoWolAuOT \]5`A,\:U@P&"QVjbiU,|RF]p)/rB:bfL B9ҝs/-*'F]& y_U6F)D;MMA&[ %QM49 c4FΚqFn~zMu RjU`pU-(=s~hP5w(; ,z]qz^QyϺ)]KI e+5?1>ety/q|uQ]`zV,{)VmM;^D񝧮\Ϯn(R^\fz){zÁl=E]J;eȦ,֕Wb. \`MƕXkŕԚw%f*[WWJT RW`IF]n{[isu4UW_`9$ts8LCQWm+URV]}!ŕ Rz(*S`U+DrA'^asèz7aT=]+ѪǾz*{t0r8u|UDBuń!Ky0*䡨L{t4UW_\Ô8epz<3Ձ7~oOr$yeWgGkOMbEd<-07N~ԙAFޑV`pa4+AZIS ',$p/C9SF&S =MUZdVd*\/†J0lxcX &b&/|MB_]KxqʎlMt^ޱx"{Sm'oVF,/ܙ?Hg- :*_oSJn0%NɲDqY4P@* 2piIFW )hj=d bm1D[ -hA n -I<]L @ءi2T컟&S]O#ڦmA bm1D[ -hA bm1D[ u%>+QS:YBix TiOB R^덽Ù8M&Ě)Yۍ=q!I&_/r0~Ek.GCu_8 &|jS( i ;ϔJ#0w~1&kLj4s>ouE?kc;K׽9P ^Ǯ)sZVG'$:''J*prKpy+8(NfǾc;ԝ|7+[*zpf+5dPPhQwx׺ ޣqn'G߮qKHB*GEڙyNuyv3⿥wף>1%~K݋vnEjGeˏ7n! F- . #3,QCK3g1^f<1/./F ;*#G?Q>V FFHX ?C_='vNܛۛwSޖ_;?]]/>yٿʱO|ߟS_Z|/s"cIy`$?i)ojho<4װЦɧ'|q4 qBOT!1"y7nA"|p4cc,vբ.fpPyӬ ;c0zm~hc!~VDhͅW}Y?Ï9oH?S:'l\ mA`HtvrR$b#7hr2MխQޑ&,JQc=$Oǁ#7F8#ZJ2o[Te(h2%ɤ NxMF6qt}!.\i'f#dMl` ~J|VԞGXMvFXyi5V2\)"5`BhpM aEV'0J@fk +ڌQ/X(󠨶"WQcg)1dKJ`,XJ[HHT!grr@!+]%#/|5B c.(\ԶphR`$G?yfO| "nj\TEƜOI*o$Oq頓qAtW!=<52q"pq4@P"I25V M#cvbjwV S\EEmMU^8t4@я9h%sov%tx|8kz=c's3n -9Hs7TIכ\éyID:%i*4$ {!t$r`)Ĕl ']LLf V ʰ׎t$"i 2-\aP#t"8MMZDAi%wp^>O4'[OS}=pL+!IUEB0% :QO5T9!d>jT3JR9!u*$XJI% TӚ1rVk(e'1x.u!ouVuLu':f_]Pn{1=siF568 ,M.$LP =Mb`"|L+ݰn{aʸwֆVk UT;-(񊠍Gj D%L{.hBxw =gAZkelVƌHhE$Ae k2Jxh(eUP1@#Ơن8 /{DU s,PFXaN }^Cﮭ}+D98ƾwSc!x` Q*0@i5 E֩@QN%ȫ6"5z{&FQ4eKVmS%Πؙ9ovd`rkw3r[_JTMBjvơEHa" ;ْ֢;}kU枯?}lTx1pha^vIˀ9=0D bTmAA$Qf$EC@'"4eN:YgK+?M>_|R09RA0LJa OXPM.:45Ɂs Zh[fG&u ҄yj vbv!9[Í?ώp꒳xN8x'|烛N> ۧw9z7#'wo^mF-27fTT8B()#eyq)g3CĶ;G 0~Q㳝*?3=:/߾T^목t'GPtڻ&'Mz0!5\ƈףm (?ƏUb 4q"3{Wp¡ccrM "mKж!mHކ4Ju"QjU r ,( 'E2\UQNA*Dy19MxQqE]|24Y?vgCTD7)uȁDDBP",, Gd iED4 zӞin7>A]u?0?~wOs?fnı6kն#W&ECKq6P}ʇ|BNa)>OCHr"RHxb +hș̩[y[XۙG:-o%(at\` 54r4K!Dȡ*7F+UOzA5 lJp߫\/ 3/fKzڦQX[#g2JhpSY]["G+^8~i5ifvv;k.0P߈$* *Gn7'";k3n 8`@H_B-و>qDyDYSFa' Aa;&P`& *t *h"=n- 8ĬHz20qTS挅G22Ӣn_AbiVyV M_yM])w4L1YnV ǫ JK{Y " (붍 ޱZfgm^T,sPA1<.UJB|6">xk?Zu^Oܟc%^Ut׬Pl*ZRYTLLAonAߡywڔHQk,;IsQ7\rbrgT‰7wQ(F리͌$THR`Y Ū0 >R]:qr+  wF|%'ǫrъ|[2~NWJ}+3v<#|Z~^_TJksDlQM*lt!+˟S*ELBE/:?ED!0ׂwG ٫Vp{_|$HűkC&9 xQ;l^Qز`$7ןmNJD()(*JfP !!5D@M(J4P+qDQuHOJ^ -HĪ.T堼.{x hRS/cd^|66+ݼ/R0hx9㔮7QǼo=?&|w^Kz /a*Xlߛm,ԬKYzZH  C Uo=05%ָ y`h"Hͣh ;e!#zD{]:1<24Co JAW 01ҾE)xXz'k(CZ|TESDhC!`D2$3zq\}֢JL=+ n9Aq೾ Zl]4ڨ}]]6]6R VNmcSnk? FGȈC<y\>OiV(Gf',?’2 GudWydQeA*9[RT;# [̊p>\*n[|`&Î9Xh0Q,ASmn+c=Љ_4YT-NQIiix  Ǵ#LJ>sٚ޿_k4|.B~afMW׽+u+u9.h%8DZI E(aűkI<r ^jtpI.lh[/iYG^M[Cb5(z'ppQƯrSBF"x!NaΉ ,PtS)x(et9^tΟΐE^ٍK>r Ŕ!݅t@**-Fף`#u&Ԝә.L?M|j5T0\t%Ꟑ9cUK MWe>z7s{0q5G^2 >9Mt/ xBxUKO{׿Y " (L?*@-`l[Bsgů0[d~]<_T}krlu>],̸F懕ӕ\ލwz*l# D ȶ4iߦb<0{#Kk%ъ#v?2a'0Lmd 0OyEIm{MW'];Dm8_|?ZW"^a|TS;(*Ipާagn9LW 0 K3l=c\;-RK6</Z#杼Pb9IE_oE5yq8f d<p<\pdUr40Man9&.3J(GB)RP{۽.q ћ鉵jyXXwvo۳Ok RfMWzSu_jR{)&Cu9\"W4D9\9l9wq%[5wFs4cJ#P [@ ޵q$v~T =xeɐmpUp$Ґ)e5UuuUeT  >pBpO1THV`E"k;B% PoH<ϓs0 ^*u:ȿqs9Ͷ{!fv9*{8_I_8vaBRNIt.&ybRDd1L8 W'SU8**%(,XZ"9P* 9XV[\.it*Ř[d*/0#g;ߋ^&E1lO.F6E`W9¨Ŋ[-e^EUy[ZAq&$@`KUsn9?+ rVۣd#gGz nFJn3%C]'g~>8f!nQ}6Ҧ^r o1P/MbPsK2#3FgJMDF)#"9-! u5bڛ\ϔDiJf JJS^+Vt"#D, GAcr8>\feduH.B ۔JJ3ք P \sBrALajTi'="-WGCHL!zIO3Ƽ1WUb TWފcBG6$]q!I_t.иZb5]#r: "8(INKcna>>Mn^uwjSB 5y!yDg#Dx?J(zMn G<>;ͫ(|J׳Ģ8~=u?ƫւd9K8oՒ͛KC'։ߩߣqn'׏6]8Gg7P^9~zeA<]Ks}}»Upv`Bsa i4<;o斛]qRx ;[ c$IYH k=fu|@F qw}9G# Syi Zr/ߌ÷ù$ڀ']7H܎ MfM,p>.*6<ĕ#NϑI>~9|wN)ӧoP "\ǨwP~y Gpg|CS^jhkehSø J>rǸ_WP-""D#x wmctM<]u̖T5Ii$ˡ, ۝6 ?s?k+A݈@+鬟oH6ꜰr-L2Z\z:,."4bHFne:{#$jV=$G~ Xaar#ZJ2o Xh2ʘ%ɤ N3&7 d5񠃐 |VԑG`]wF`=tV(ʏ=u.$x(2A ꝫNBekVy^FDQsAQOS%Rg-ךRO!CL$]T/  & *N#+ñe,|]5Z{*!Ձ  ,./IP&s*FVƆTd>6T^w!]ªգemEY]@m4=3еl{u R> on~U쳙N )"J*WڤX%#v4JS}>̑!'ehb"%c@H3&!:ʹɨD8/Uhc"1MdҼuPHY"!(OH1Yx9;n7%/fW/#)U`g9j{&}k)oߡ-ue8u?.얮Zx'EýSIk eꪩ@h{z~dOmіȭu!լ;)jN7Ki{^iy=L1wYdvhI|tKD'7\%h7D]/ۦFu޵BJyv:d)4ҡbls \]P Q+!%"J"X+"*rh7b=rOV>uDcVrz \dd'<j)頓q}-Ci5BnY5zk֞1:L[=|b6iZ#Y|!nU=:::R0 lGW_ΦP">DzI W X){Tohl 7q#JJ`:sv}eafaͫןag>' ێf:fY&-of{q3uJUh|IC>K8ӑˁSVHXgt11˚@a9ґH2>9-h+Dːb@)E/m"g;ƳYm (Yh'UG  $DIL&.,]b8"cKnZ'"hQΤ09ThK\<22 J$hU`"-`zR@+PքHdi/8ƜtǐD l}.k>( ĈX$bP( KDK^"VFSB4[NM E k2JxhցeUyy(l{֮zN;zN{(xYl=xMD} ف M~qp^⬂X" JXFVLk^U$O'=q`(˧O[/,>X)(^YBќG)v-EGJE!:|sa=76pٽhPrQ>O5Zks J;|K?KcK6>x5Ǖ9`9.fNa$\Au0b{pY/-osV\Tg<+D*T+OL3Ub"̩۠{4r.]X>bGzA@ PRj"ȝ "tJT@A!T nV<>bLwW H}['2zq/ϓBAf_ݺ,ӶzJr CCA=IMeʞcs~kx:E DKF"ʟTx9;$j >Y+7r{rƪLPnY<\) 6ilUӛ2V|6ɧJoz;Z0?P$E@DS_|eBHO9QA+?:͕Nl2Z?F,kޚ9+4e\lvb2J< ...Sxh^cv$¦UGef)(}B0?GD7i׮/&jq@}RTe8+bt<'FGzb4LP!)DCP$`[ kzϩR,h_w,νnKdJk$xZWl5_6B7S/M^ͣ͐I<]l)g UcĵJ"2 9aICS'SQ\r7&Άx>7~ qaE.qA .wQgXZR ;ޤxVq]/y-7ZΛܨx9Gmx()=KE)$&j uU$\ػ  ^g :Ea4f)G0V&CHzk8hi! a ս뱞P%mqbsu׻X7N*`Hr6(KF"1DonT[FhP,)FHBh $UYL2v!5@;@#8 Ͽ\09-'we&a{bsX,"g8 > 1ItSXC,хq:}ir]mO_ Ȼē wnb["8g\RS&e|G#zW/FcR_`UkfHV9}d#y2wksq H76iE9>~-u/?U׊d:J4'M۫5uG0SŁ"ntu9kG 6/i8-P9`wn?랝OV]0Tt9 >-Is'PwqY7by7jy7-ĄAh(|]ztgcw,UսrYK)qrFс5R:V4}6WKwOȝFkߍsS7+O7Q,qW&^L׺dS';ҡ1 ?NwG8m.Zbţj3ynrHۏNwNooO_} {O߽}E{_轌1:FDXxz_t?m5MuMͺ65ߠ_M>~v_ŭ~TKa@`]⍻OZH9ʕjs("zsj7QCE  6)xPOzjN56'tf`O{Ṗ<k_uKMml{Sןiqr^ƨS r(C@QdrFAk<ϣxDQs|AQOSMy^Zyd@d1{e@~M2EDŽP6Yi|go60$^Fk}J[H#Heհ@*Zg8HN4.2t|UɎs8y;'[ e(k-Z=Zu t{၁ͶEZ;BZN -OCs٣/'T⳻hN1(T- kGZ[oNltAw3 ;]vW놩~ϓoFΤ_zgً]\ԗ;a4QA?t&rIuEfcjp9z1|ӄS~_GWVTַ? u|Rg& r張SU@뫠AUh$1ԃLOn{5oN `sFoM%16A 1{ᕊY̎{1 lv4kNۚtlPZBz&iױšlΨ)mVx OGJ7OXa]duؒ9ldW7Aꏳ h"MGˢB!x(dZn:9 F3xtn}Y٭ipϨzc\dw1[s`?̕d4i.x3W0欝VzCKdJ&E,0 m%ɑYvU~W K:g0B*cMӦYDlXBqx;b?g64OZ3Q{EDGR?Q@z4*TU Q9QYV2۬mXc_vdEb=OwJS#:0r.n=Ee}+ŅG<& *K+ u.TAE"Ykc!#6LfѴ)qCI L0*'۞GN֓ _bbf{wow~Z}{h|b&ۃ=iyTa\eUq tiId<r܁O(3F5,uL L*٪Q)U9^7+M#YH'X?-T|hoe=iOFß\3rXcwF*i6O`CPpɒg4W❢gBOSp^Ϸx˚_2x7Pwۖ(Yٻe XldR.T1*XFayjē5G]QI4Eucс4\8v.=wkOn@,5c݌u7rΥ86E` SצkġԵ)Zumm]RCRWE*ZHDFue_NU?8\t^ys^:Q0:E/4Sv]떝*-͆rJs5{59* vj^.KyR5{/)%f=ԬiC_=+XXv0-E];+z/)֭ Օ*>uUUvUr0ЪG]IH`"EW_R ilz)ˤ7߿=\,um{F5V: Q<'Ge4Sr$4A n%.yKC,~P t`d\\: o r+3s+}kqЇd+e=7VA"}АhCJ&D!Lw AۣњlxbBfQTU HAPr+e:G\9sY& qs|^ejqivwr|kqZGIY%s,iMH{7<6 rk軛G+^ki<PmЯWz%;1ͧ075iF%-S³94 L=é2=p%#yg8\p"Ј$R0*d۪Q?p$ Ut[t: JޅT Y8%ɕ~كka*<_$~Lu\_(q΃?0eV ǫoW^KA>{=hJA(Fdʯ>̓E}X]"3_80 rOκja;?b}uvd<v>k6m/Gק:} fz$h| +Ȧw:hjIVwF8,^ɄJs;wE,9;$Ͳsc4rK0^3^1{|;{ ЫoXzxe :h.?\:We,z֒Z6n2j[^ ʾrd6Bཥ{Joڼxߎ\= KaŽEr·um<{ p=wbO,Rq-kݓSe%+e&v6QǑj9^v6S*-EG_VK:QԬq8V T~zyx8`V=z`ZkI訥 p#LqFh"2ksJIOzK3y8~an}{̍aSaw=JnvP^ "l:y[uUSTJ'xNlE`-φ>kΆHkO>HyjzWOi U?Uhxe gr[aDl h8 2&rV?p? VEi+7T;0=E922`+ ѲXg$oZvG!FA{W15f~*ԷGs<l)L@aUZ"2 9aICS'SQ\rJǶۍp0 w?E\xǜ>7QZZ7WB%"rsHdǼ?2̮}?}[b0qeu=eҿԊG⁑O'2p&KkVR$ qIM]w$*nd5kfHV9}d#y2Ok"Ѩ7Ҁ 6FC4zշ61[S$bF1\$ BW&Ξ %P̌6בz(U৅ vwl:(\ߣ;3p6 ~\.;^7=m. <<9;?lUE!"]s !ޑ7X(w7+ٯfEh>*Mslכ6ZFUkwn)E~9?2_ZU֞zK\=}iRߵϦ'y uhy}<7\1mAmvF&(u1Q"Wmj&hP`Eﺿd'd;uOV? `sFoM˜B JŬMtZf=M6`Gۂxcq(-a]0Eo+udL v_Oq>Wܹkpأ7BGqϨ ,:zE\u6>Fz"ڜ}|=>$hbrQ[Ϭjh#wY4(ddJ`t0;qxpfI6R_^~fvqwK^R].62\gu9k Z #PDNJOt,5f(c$GfETJ ٰ6!LZ \1QZVQZBqp5}+͓Ck[=a^S_^1HybVy7JYdemP@ '5K6sYfIYJ 0Řj59S6QςL[gI<\Hؤs>m2nG)ob ͌Cm!TG7e㪣(➩OY\<=?M>n)pDQeGpt^B k:ReG.J-G@) C|L (fSj5 L&wY8s̢-v5qv[05Tvq*Vz#m3^IQFB+WNp8YHA^2gGo%.JUC&dȁ(ѐIȢ(lpaJ,S.+ZMu stSǡQUE6ざv `&ԁ! ,h%/=xiɷ49dE`.I1m|cT pL )D Qs#ђ^pRktVgE/,uJjPh*E.:xQճX=CY۲l7FJ$W>^X3?*4\O'9N*`/G7p؄Ȍ˂g"PԴ{[ e*jxvE-dK&# () 1X/cF ,Bk A&A€ʀYrL0͕zg aDnӑY3s=jaS2W x0y$-q\fW>4dA:rZgv41/[$Uy(O^^};OX@pJj?Qe ~-؂GUٚm] 91C5Ĝ^l zkH] WqrK+IJmhǤS<* T5J7 1jN?mN =&?_ޚMwv7e!Eo1qU },iedI0x {{'hq7skGޖoe>R|fO%c®ln;tE;(ʀ)"`DV4ZH  تpP4;ZNYޣ áthKقsٟ5ΒE6ϛ[2}Ҝq.zeXP )1 pIJsԨQ۬X,&(E*R`+s G?~I;xvP a%aB`_*zu'ی~5Qгmg2 # D^ s> r!G`sPUkJ%_K&0;$h)SR=Ƣ-C6.sˋSOUGՓ9RK)b;R JKj1 M34֢t%xwTrnej겓uǺF_xi0oh|tot~~smvt kC2W/-vݓveAMyXh9"f-ه)bYcp8XE¡U> \コGWp4hઅZ:\WZ#;v X5 Ze8p^+_gFΛd'ݭS4:.?Yq8ɘ6J$nC/l<8ɟ)_)P&L9 fW}E/8DifxM'#Evia+EFqV.+no?V|*#koН5ڋUS77?S,ު(&-" r:YݪTC#﶑ ]tcRt"REۮNd jYв۴싎6KN{m՜Qǡ9{ fѻV 2 mSxd",A{د `u=m`.Jw`;BbAbo huDpp4puXEԡU2> \ipZ#GW-ܠ8ڝ'lW-J5WVpTX0[*YUL>`1592*lrЮ8XUnc= u6L.|Ʀu5o~beO_wg;Foi֏t[zm^<5JxI F (!+ %"g2  URRԢRMG}G]T(ʖj9MJV &y& DсdLt-TIA*2aDT ]ZFG5fV %ͱ+ G@9!wNCΚW9YT_nkҭf~Q͇$f(]}o}C'gs]ʟZמ&v'ۮ¡׏U+*?*cKuxv6"yCq$US #:kAB+9uEF|-"ʖoB(5C ֗QDW8LRB R[8#zrýzƦX XS,kYa t?H_;6ͳtÿyHzt>s.YR*3("cuTj%(VBLm2FuL ?R jltPW{2Ī!6Cyq2.KPP[8ԆS/Yf=b"dRlN4w@J>yêDR}lΧ˖Vِgi[&N1qd KEfSꌇyЯ b)"ΈhDq]ڕbC%ihfC VK2 NLTj 讈C偵1Y0*V)-1`9 ֎4x⁉:#bp^P3qS#T,!uv%nm[\. v~3<%$>]o}ϑϮ6A;r/ՎzAV`ýΆ;҆鐠6 _J  lV;E,TS@Z@%oQOEnQ~VIB(r b1x5j(G,@%(PBI"UWP*Kj_X s2)#d(d5-*_H' ;GW8/yy HJR 'YIJ8/BnΚ~M68}طgWKFVrm͕ĎZQ6՝t ,?$G=(PʉEAfAk`Ds# 4Jh` ͂I*A(VXqZ_$dE *UHDucؔMLTA,YRBu6kBU#z<;dF0éhg}ߚzjkrܭoc{V֊9hn<+\>C?Xq|+sWşWZC E.dSR{-V ]T>R֤l˟jq!T:gUTX+eEJػcgLرTm;pվԜ2"o9!<ihs`B UJ[BGU::@Sb!B.ːfnqzonpwlS?8}i}VߧGw4⢯ яK \RR"D)5j6+K e(T, ƜBяF#ރv{d:B;H=IY[fH\ҘX$D x= .pJSݣW!6w|@?Xu7#|q%*=.+fp痳r1iυsUM:# IP Yod E6'[ro}˽IB<*T9c(B |%L62+g $-=X{DysB''?y}Njm5O+tZ:\wsVIoVL"xz w]zi즫<m93@Dt`N8v>]}I\.&E< ꝰ)eܐ>dj2EPTdd:>1QiGuƢŪзXM:r$0$f{qousF3QBUg%XU|1Gbf weCtlH&gg aYgZ8B;rCɞMrAO0E2*pU%ZRTS$v?8fUg2W`WR g:ZYP} u>Yko㺑2[ZO) ̊.+rS$`9ڏq^d]@g^o_I50wC*M.f"xZ8,"؄d> E0`O0,"3 1Td+^0m`]ON:J#57lCw'2iC% z xPyZ)i5xtD8\m}WOÍ!SŽ< $͎+fIuG0lXoýƓ3{̜ <0InدlXcf`Mxg3eU7Sq`:dz);yaȁN }EmSn''|/@.OX0 !$ ,rüFU6Jl >ToJjID!֜2u[S喣N"XTr,oUr~K=<~Ft\+Q/CKB9grYuzLyux{Iޫ߂M|Y0kxU[E: y佩?F؊>|sMD0LJ'ڧ_Ay׷4r)}F-Gv;4*Y)EC@(Kt%Ys(#gWoWE2M 3|ZhBxΆޏ <=?M^l_AlYF7̴~i?j~m%zofO~m ̿^?3 Vz~ךިA`ZIॻEw7,HOF^f'| ilE>^[5ѳkU Ui|5TTkne+߮7P6/P<8+j@#A*E"GT WaZ}gK`RGTQJmJ%&zGG}S]3ξ@*(] l`@}0Z=l :J:Xng=gS#1HVKW)C5Б8UJÉ= 8`j;.ѩ"RE S(uh.A-y3d&#QHYu޲D365f2b=6M&+%"i6rv {hu0N_Pi:k&)2a|=Mچ{wpKbnld2̙2}xiÍ[5yNjlڞI@bxݹ4jGgLX-zn~*uoWޯf^Qa>n~H ?ݞ>x~4ۙx&2}t+oݴšן8i׌xXOD X) SH*wx/ps))6]qAiA쵿puI>5\#r禼Fa@g>0kܧ4( C=6.MԽ^}r?i|O7yL҇ȣ}U+0s 0 &0un]͞T8N[lBX.*L tiu!-}8EV$;Қ']mN! ƔW#Y1/]t U+Б8u=?[d'l+#a2u7z|]x~4E-s?eboMWy]ߺ#nן5iaLuL75mz"򚶇 ;0v>R_܊m~$ۜ2Mw 2}AŕmF8(XQb ``)$ŤO Yݬ}O-rFhSy )m *ϼuùB:-h`Q^Eme^`WG&g}#uՠ9:0\iIoN}n8ob6bvWǂT6|n@՗1"rXGRBӅÕV4cV"ZP P'wAZSΰkv-Ȇ\ ejf!ɧcc>dbbSE3GoÎYD HH!Fj`#c6D"T^C!ItD <* 8C*2I(P{+%R.!"& &xOμbyJw viuBxKYe& U(D)hd!\pXh޹WV(GkTV&zVK%+2 ,3AI֌ٮQQta6W̺]xP]x/4fz|&}m~}WoTf0}70 )@(`G%&AaZV1(mEF6lC'ZK(Ij yI=t$LEt!"'Ufێ~Ÿ<;NEk xJ#QT"00%AS1"֊`6΋`#88f`!2 hG]rRyW0}904&>Fv}CRǾgֈhĢ8ю1$L;@ F{C 3TmGh@޺sZad`(QHLLN43[XA`Iۑ$5`Z\6rkj z3ef(^,z (p  <[ȑ1LuF lF,vQE/B/>~*,cY\YVHWT\ 37dE؛ڛK@'.W/jGzEm?jڍ~= 4D}6*4ؐqD)d2?ǂc}-6'+D{WHe T̢>Rl2+ŅG|N@j+PD* yZNw"=J[ tЯJRLeT$m/^zQ{1F[Yͺvrǿ3E2XL e/ otmNSi$cɺ-YgyTX*uYCE]$ukQ9ȧ`+%GkwP:D&T]!lm1R(C{ ((E}a & 8={%gYnظc Ӗ]뉩]婹L.d! WXcwF*^Ci&U\tu4Aco$(Ꞧ-سz[ʃ`Te'XZ vo)jWU'fYk2E;ah %\@D#X2OM$FYR1t:!ږFŏWsrN,iVdrVqvnzۚx؀p=NΆ I{nJ JNh-Sq"9% 42ɨSp4oF{A_v6B;0pF\ `_ivW>=;iOin d~)6buu&(OZH&!sc >)BcИ#9BZgH%*{SRdBJ#z*N*e1[JRigd2-(w$dqLRQqgrjLT`:#Dy̧k(,WYɺb]'Z2cudX7z}zqiVt-zFw[bW4o{t>F4k @J'ph5t# yhXȍc S%N*OT*Lq3'm9 rn1X$/ml۸BDXT>x3fo }¸igj} 癨pLɝQI^ uVꡮ_cU *k>cy6J*[) E暠kPF&ۑ = ,衉'+CrM#U H/!}Tr`RBbt;epRM,9ƵY# '4ljt&ra\׆3>I-ʇ<99^ MUG>!ћdٛ3}~niO*9k.>]O'׿?y1fՁTzel*nVs2GLTOkۂ.pl>(}siPN'X]'0ץ_$fWB>4|mè/ga}>ׯY+MG"m ]n{ 9hg7"B{3hOlCM|뎩G꽆p;Օ&Խ\д(.݅qզ hfuy>>܍)&ueξK~,i.0:ĽEΝ4띴;.;I p>=Ll.ܱH GM:n-_k݌ƚ2[_QFKwݰ{ΰTiYW!`'#z(J,;hd.w3{tz4YZ(OɱZP :T0zg\DfmtNI2aѪƂh>/[ {>nevޚ]ؗvk+5H^jQ1RC`T8ޕo{:D?"<4NPжq&&R&ʘd(cct1;FK$WZh+B&oSdic*R~bBxϦ|YIjP?p|/?#?. w>8ݹ`ǟi~?~u;ֆHəX)簋V2$$wZ;k8W!EvSW@;5д؈s,t=v1-jX~Zn?ƹ늃p-6thw\&Oo''/םy[`g'2pbz:[;`T_D؛+ve\%B}36N]LD.O'~H^\A`+ً*.ɗWUZpU4l$îƽ:p)W\ܣֵƿ1zX-i#xb ] X[PVAwQ/]rsşJ| ]C<ԟ4tuh&VGD b4z4sz.pg+nJ޸qRIoLr6]z0_vevܓL%Tq6et>RZ@r-Y;O -"ۅc#<мNOyF]Lbd]d^Jl\[I#-Y]GȽ$qw!N6p:!iNΰ]oNSR\G/:9ʶɰ༒"+5LI F<)g^]ym Ɖq%K.!>J|bmB e|I$E]ωS1,izwxw 6WB!gRl7JKWCJkr9CCe6XrNxK28Q)\} IT&+-DI"dz/_o}TR1!)q %Q"ȕ `R*r*slZ5t-TX YTc#CFG"E+1xϴIFcPfr$?%z'.j%vSS+&IHhH&!* ohkCxѰ*%}*Ś]13k^`62By:(=qYc}"*'-QsvZ.xmc"vw^x ˡGQX+'mVh +/֨GpۨCl!T85LTK1@-Xd/ "l*딅r  H%5(cX21$ ә_I.r^1CM1S 0rƋ  3G^L.)0`_1>a"H9d| ce$\x`fZS !unegV rLA˃*tCY EbȃcPmVa!uP4 CKHrW4u2$Jx3pYf!0#k?^a2pJꬆ7LjOPk w9A79ᩤ1 (:k`[ժXTLfrb81@6(2\E?"7+d34,S3d+1rfLz҂@Z9,%0!.'7!j(g[ۈ;* D`ȇvGa02?%( a"e21IA1YJxD _X`z$]D8B{Sa*3((0s8e^a@ : AK/u2 zJRѕp0 ]JJh5{PRR']n8D**Z@L.H8[VScr;`ɗUXzaBzPr 1!DL W$X+Bi ,Ѹ~$/.D,/x&7VDlh ',:;**@ɉ⡫;͊c-&sJ yYI J ;>,Ƈ`as-׭ȵw!N‚.&jH \{#@ xtm^b<IrC*f7ȻK}Y]P>B(k)d$e@{zG:&oC@FL3q!2jxrI'#6Q'kFV(-5*_84'e#s4PqإF@uFu39(k $A~2C #RFdCQasWuCϢb! ӉTIXs?pe:0M`Y5 3Ko^5RR\WoEeНZz7j[ 3C8]$/q5$oQBeڋ!K-AhL1G#Knd$>j9FZ6l!& _Ko".ń@ZĔp2"ðB% "!J؄t^Q튄 S q1Հ]mj^8-R;ᢡ@]fTz U}/[Wg%W*u@koL>kۋ.kv7HcY?*ͺe[U/ <~9L ^ qQZ*3 zEe ڣJg~_V9TiVǃ%ת7woZ&V]ɺt*ԹbS/{B ԩ"q?"qWw9W7u?sHS \s%5C4-)l+;P+qc /Zi?9?ԗoVD?DŽ>Hp^VHO9tх tC nIIW-aQ{)-J;41 'LHUpa?38a$#qHF08a$#qHF08a$#qHF08a$#qHF08a$#qHF08a$#qHWaTL1)_ lՋ06q})FH˙}&%a|FŌu ՚hg ?7'@~tN}{ߥ}X|<*,::^M&4F;.Õ-,ǑhG?r4 k>DiYavюW{[:+|uu~nT =`zu~\^ oz؟R]ʸQ cݠ>lmat9f 2km%Yo^d`Y$ gau\6rf3ӭy] uO,NӋ\Vdť3l9F_,a|i֝zn;XӫpՏ}qz_ iv~wf}l_gqN/N I*[sfo^YՎZ$γdypgH)r%`V|oJ'ڻXo p@FcxZ3='n^m(Vn㧟9^ƽ¼|sY`r7/mykuXOzvTؤθ/U5?zz͕Tr<k.iC{xĻ'=wO{xĻ'=wO{xĻ'=wO{xĻ'=wO{xĻ'=wO{xĻ'L ɚC -nW\5{rw3T =R+36zՀc ?t1Mt+- ]qgd~y8;he N19_kK~7z~Ggl'qwY sb} &H9O/'i޵rpeyꃊ|R Ɲ_U=BPdն2pw2;:˗7%Wmszf* 8/rmኑGxty12q^u c&Gt sD%_'x oNwm>6z3}X"VsSO3_I.nrm0Ry&VL#(XW%BBYo\ i+<8gv8˄>]y*FWgbi(MZӄC s"TȜ(r=6LgzHMo#q>xVY&(vHi:d ZuX7 sqƻm/!8_\h 8Vf n&}j֡Nޟ<˓h0NZyۛΚKh >B+1k+ݕE(F*-K<C8q8n/6͏bsսPmq&3XVTAej&b<Ԋ(Vf|[z1kn^W\vߢ\~LKS^֕J;?;w}/n]{ zmneJ]Nݪ^R*VEݪ[unUԭUQ*VEݪ[unUԭUQ*VEݪ[unUԭUQ*VEݪ[unUԭUQ*VEݪ[unUԭUQ*VEݪ[unU[N| l2]7ϡ|MYtZTVHw;oK"M2e+Ty-Uijkr{&ev"dtBZ+HL;2Smnb*!Ɔ,,Yyc4xNlYvJiC3tc)wjqjpLAi1/Uۦ4NHpGZLDkp" E<~-x4q[N%r:P؋/)9EޓWi2˖ZIXŝI:b: wVm"u~5w\!E_aޅsr٩>a+[TNXi`Tv,~FbU\A CWxrh$naued;)K*\9ՙY 2Em֣cl1?Mm/RC}p⳯=~zYh`v; HҎ N嵨עZT^kQy-*E嵨עZT^kQy-*E嵨עZT^kQy-*E嵨עZT^kQy-*E嵨עZT^kQy-*E嵨עZT^kZw nⷓ7NV-jՆpߩ<_?࿚8;GlC\o_ g4cxmH)-q_#gHecsqy9]ʽ[RtT&`}`ޏ@ٕ+\,$;`Uѐp-hq8bϓմѕ:;Ǡb6Qxlpq#~D18ew%BgA~PYyg lV9t[ƁgofsǤy2zS\  َw7R"PFʒ-U!}5m=O'&f[T_i8sNe:k׹E1*pTP?:,^&7.b%'.5o;M۷ILEG)UƠIXh4qC?B l2T'M']l/.I|vznVov|55a1[բU=0* ?bя)];ZȽgCY4麋Ѐۨ-67\F gw8vv<{s:3;٭aMO>F!VJ׵9-ܺXRO=\uƼ{X<{T/ Uޡ/QnD󖍟dœ",_$>4S9f ]W*.*:.s/U{y={|\{ ;tx>?x,7F W!iNt̷nX^]O ,5jĭE hLK׻pjBAqߗiYn:jn15յ`.1WZt1[HF_kmXe/I;R^7Q.%яjk>d>"+[=CRDR 5(gDf5=TWW BO55=hQ.C*JDp .\<@ x"&ځ@$p%I.jlA1 % aG슿%zNr7Ah8"nκf 5?#v"t//p|g?{w7o:wop.B$;U_#"i_4M5͛FocD}s IjQuQ7^dPeXqrl3#A{Nsb=+^kt9._&|P*@mb%0 Zziy)$`sJ[T7~h.!{lfx}|.%j-ek,Ҝ' b$~dzL'84Ţr gu\!%K˻H^yHLf ڲ3H$2K9ۉЌFPeZ&MsQvCmLOv]1=%,_wUI9 YH4H^HAlUϋ&v@K_ JI{C[V~>;f GH@2=r$yFFO0]W2Υ :hAϫ(U)W20^L+MxbM+5%-tb2 n(3.=Oݨ |KQ7wG 1 b*A~YNN=e@ (7dagի_xU|M/y'<΍Ϧ >@8~]8?CU#x)?TYE6/_rOm5Kh\(GuhR<&<"YP$F"K˞@_J6]5@jtBKN(uPi9H$mHcx!JB%ɍL:bQId\J \Hl63-yOcYFjw^ qRDW(DB1IoH(FgBhJV#+Y ՛s;czxӬnm㝷0UR1և񖖻>wUX|̾ 8lM[nX񲏥XuJVĴ=ʹuNrȤnGHP"v(nGY6{HNmph3=H!mP'c0pBQ鬷(xxHT'F9iy>xfeS5\fj;uls[E%^T.]t%jV8j9\p4wx˃pHk}Ք MK8`W7;sY粸.},ZL#mB-z&E'g9=f\kgC,"RJ+),EKceY/erݪ~-iWUoJ)[M1,jc$B%-irKkJP!X2FN M|05u1Ba ;\i=NFĦ6&BϿ:y%"P*_t6mnb5s5 wĀײr*!C.i2;q0<;)ZD$D-* E $Ɨt,Vn9]dL\Q6HrMcCXMețqh_<x Pg02X9IA#y,M >0"6BWUQ TS7q8hDR(U%\`ùgG.=j&!DjRiKͼ"ɖmZ=(~~~q!XI@ ?n/֮Gds9#{DaR C`hIo]/ʃa||w}0e~=%D|,nD$9E%.9&Ny8r1}hw~:@ W UdKسWYIxP)R 9"(<*lr~uƹe,m8W\*A 4G3ͮ.˳ugU'L'0Is#L}R{~Q-7}(BU|N-;n ^ B6]k[r}KMͰ,+aes/|8];z|fߝl{8+6VՍj%19_1>Bt+#~Lh1Tܴݰ8m9X|~y3?t!]W^*xv <e%"οR` G: Zw\-Y,a[U~>]o~woϳ7Y29s ΅{@hKxW~y(uMST޼inCM>km+GEȧoG `{vэ3_vhb#%;d}W[nKvlJ7A>K*V Z>_7}=x#z㾥źتķ/ՏX_zU:MJ8_$?@ ic b!O`Lk\P8Y?L-k)oQ& %xkXFEa1:Ee"ZPD$0QFYPE%J?}ώ,C!9kL0 iJaB~ޛ c2 hRR"q`WԳЫu8[=M5vC;Ǐc~*ܜ.[Dj Aw>=;- BvF&Į(Rh t{ż6(fH\;K/͡E b%~L'ON4^$Ph$JBg5osfE=~s.MePHnx_oGt%:tR?.h<:hxgG%87}d<o/&Wc3g'~j{CEWq̛˜68޹'ܩO>0=]Lq:?=k35;v9U(+j3:~[ff]FHgX3g5_}v._}Pw*u~ =cﻭ.]W˛?_|a}W3<ذ@z;z?ԟt/v;:w9lo8y&Qa7їtej~\W._O4H3* ǷEv@BCC !*Es^zДgd\ !+X8 S*u)X]RRR>еNۙ͹eUV̗Q.U؋9XpuLY[N姲BkX1Db;1Gxg'H١ (tBw  mxhpɴ0ɢ\*Cdl2#08KCX0_[_7@!QH~g8/uUu }'U2"e}[eUvEg#ObۨfB4[AZ4>Vv̹iO虖f!oN:|<~u!ŀhgU)*ߝd' H5=#M""79DcŒ+Tj,JQx>CRPHlW66[iu@, R9w#c;Ү6QP,X,)_TfH|=f_fi1x˟#vѰjGi,ꌤ!#2Z!4QlL kb/f%4Ҫ&Kd-dBv(&zITDr1b73n8_Zǡ6 =0m-^"v"ֈ eM"T5Wp|"Ӭr ٖc]LV %ZёcM&"eH54\LE5a3s^dL"if@%xhBbF 3Jz-cQGڳS*25a26eR뀵12ήN1`,Y4y֎8䎉uf܍Gğge(4٬<]c\.G*C11+JX$r9A y]CpQ&tʖwZǡ{}nfm[Dя)'G?N>-L};ˢ8AZԩe@H4NifHONv(!C"!&r` PL2”6sE`Y*^Y$lu3s9Ř//dq-fe":~V\qv??y,%r R`rrPމm-9,)/vbh˖:'k_}8 w:jZ2&ո73Nb]U VwH4w\xS|Aw$\ސee[fϺIVu A>vuºyY+b:mW npGI7OXZ9詓ѲleNa2c":8 > t*ڬM<џq -{]O$Ӿ.E?s)Bq1Er;&utNPbp .Cu f,̲WYD23PN* RAHQPICF#'.S3F*ML:Dʌ (6HD(>UQބf5~!N*}$қu5gA;f0zcOD15~wEf>ʛ[v_䓊ya}uZJD|銟6]ɸ@HS4u'7WMhHU!E+S) Č΄R(Ibh8b0b_>|\imN[N;PYV&6pkȍVFtYP49 nCmy텈юB5ܝr|Ů]e_~e 'kh j:wQ $s@iS$fG雧K|V5M.ϏF~^tޕF7\gޮ{y >`s2q-]睛>cy-m^a9g99i+[<+I(A%(T&IK a(@ǂ d3}nюoGgޜ6{6B;H'08 /)y|v^^h51ٙ)Ny;;:Qo! 21/njݷc19)HH. ךIgOFb#e }ZCT˜!IM:IH[q<;;YӖ”3~GώOׂOT:Ap>OD}ga;(!~­Ct1GkBe z֮?J)WeVhe~֏kk9z݂t~7 YDW.n롋CycZyW &[ zMɂo{>7X~KLף{soFooxr9ް i߶oiׇ8Ӿ91iIx?{2u̬mr]Er ]eD]8xv>2_bwr~RX"-v&[8d`AXPJIY@:R8tHNe1jK,\l[=SF$fiRHSE&ǀ"IxQJb̹G(_@繑$,O ilS݈bK|O%{%Y'0W}$fOW6DΕg2XS4ȎL'cJwJZ }$](LN2]3{e$R@2U&e'ڦ.ٍ<qۉUg]ӳ  )L ^gϛ!,%1ʅ, R2|C Bm7 0-헨 ߋY$5FIJ*b E^=!LcxH|ΛY'aYΠMwןu~v Swܸmwb)>46mw 74ʚ Lgi0f@Njsh\V.c笺8>*%1 ),~g&eϳG^{>4xyU#ggH:i4;F0 aIR4!{plTT>*$Ot>}sOZ -$%~:#XCw5#Л,_JT٩VV+c|h>R6BجˎuoҮ\gSsͭ|)W;tS\-ҾItisj)m-9z&QgpPEG{8JC9gJJu )rmZau /*r k.J98gPPZfn%PQoYOV"WJ«lY-H0)ŶgL~J5:WfSt$s܍`mTog}g.y6fk^e:=buWPkYq/i ЌR /⢳ٸԔPw]SJU>NJ;yausJ8_k0;_f竴VOe~wљonrQUڬLe,8їN O2nWמ*UԵH5]MH:ړڳ._{VnaNJ;z*灛^GӒ_)e h5*>ۥycumkoNUWRP&(;a[cױ[JǦ|WX⻓ն1|Dx5}O^ #K!5=+*{iKӖh:ʡ-QJtR(~(zESDсv)w-*^ZvbU;[ۿy )@EecQťߨ.Qfm*scW0PVגZ9{Yɶ~C4ߌo^XnJ=OjyLӷ+PcLS7T{)QD&y 0k]%\5Z)OJz/RVUYS**THWLI*e 3CW .%M|>HW\iIJKztJJhǮJZW/ ѕaJpIcB -=yg0m%ҕԬ|U,JpUcVS+@lҕJ &9UjWZ NJƮ^"]i]%9\BWV>]%Xtl W=>3ŎNW[Kj1Wv(ىŮt[U1y*1tJJh5?ubW-]= ]Q$ +i ]%7Z"OJ[ztŐD7`CW j3kgp+u2銃+q*>vp5o ]ZUB[zt%nuU &Z:] -Zz9t%I;)\**dHW /]%OpAW]tJJhOgPjK+M:pCWWUBKNJ[zt%8a˛IlrwƦ<3}P"9 TL:1[g˭UUOԯn1.uqūZKjcJ>" *R.q#yᷳ4^ntiLz}+ݛQ=pVjJ/js=Ye)R$R2 t>,o<0L.HT\P9"_zr:.vhő^meId .DaDWi\7$Kw\ Pu/UE W}xbNM%=^H'r<=ɍ2Y+gV#j a if2߷ KQF+"yb{,X=`<ɠ`4?ޣnna3ĖIF } nSa4Yn ?鍷z1;4qn&Mƈv3(Vz6߮c= 5 #.>Zλ>^}+Qdz`+ mϪ.p ܝhWTˬ{j20K-M%Oip3y8U+v Fs$wiqΑ,?\Лlz~.L:VB󞂩e˜a|ҧH xd†H7ŐiaGDQ3' lY$P@{+%R@T]p6Ouvftnۺ&(4s =6Jggkɡd{iw|y;Xx? UTuA|XiNNrgjJ`ԇ#sUw =r7M_m98F. I{&1< i!ф 4( -g8P M"c˺3r z;}Șkϩf;oG{b/)ȁDB1Ioೳ<)=Qm"6G1CȩکS-WHА}1W!UQk$ Rv;2T j%' c!B|ouo&Xu$FdX)Ө->d}vIB8޴? uRpK]_ ڬ5saDvb\ؚ2XgN!lEAJ`Hhu! )b UTC0 Eib< ik[# Ez!VTbSNzy YԓOQ~ʬ֓nPGy*Tbe'I0qrp9x2˃pHk}Ք MK؉V0>Dx⠃uctA?Yunkfuu@$UT+yBPLƸH{EKh 20VX!Xc Ir^x ԉ;1:X+e0G{rKnC\hh@ /8 ;dZ꼎F9 ۉ6;Eю-UC^.V;$Ykm6ط=y%|#Xf\MJY;@4 <[EDB&{Ds$wzn-ӷRt)XZP+1͆޴q0͢4xp$x1QoX.4?g[jwOxa1 -^e >xυrT'p@2d8(A#%Q T68͚䍳s񏃺͠!+A%yYfݪpc\b\a0&jD`B8 <+HŸj/lcs0^`FzhcD$A;B5C*2@Q*Ir|)΅cD>ЗZ蔜l}^Tș׳ŚѭpoΆ?N7y-G(kP{[$XSfʺ7[io]U:}{B1q4O[Mư8uj䏚~^ET7@FJZKod ʼӃ瀷4z9^vp|#DBK`J"B8KN"tJTģP%FtOs[Z^(zSfq8֛x8B9lߢZP' Eu~a|۫lBʹiio/־'@ߗxg,h!3_/֮wS;>➥W,欲6˶c+!SŬ杛zE=1fӻynp2En*iUȨx*sQ2ޘsSϥ*dzcP)U9<0Lrh O[Cs}&lAL!!'Sey6΁P-&'Q'5 1]Fv\VevkUB2teï%Hu cp5mPYӝקg{M^1ׄ_ymJfC՚mtڍK͈2%ql#Sg*e)A`WغK\ߖ ip뭅ώjǜc 5f <2I~Tq4ݲ@ͨ}_ad8&'͌ٲlbz3nˍa=œy[ Dt{7ơn^ %7e,i;dRX4ڒc`nDU*o BÓKMNNՉϨ>R?9~ߨ>^N $TPSbV[P15-(́!]n:vkWu5Qu]I}: O^W@)g ?N#ӆ_)m4o5Xt%|t )h%,+SX 0QHxܧ>~W97 @0Ih({N;mOuDHD@D{cYT莑FNڰ,y  VhV d}~ۢ>"{|7un`)*FHk&s+Ӗ,rN5h-),:ڭKe;R,p 8_=`嵬nprd4(MܑȬ;)RQ)@K.:nEoA7Un>@.lecb`Z!`FR0-NPD2V FSD@ @ζF B>|= *5CP¼3)<'IA\,MU&p+l : 'X@]]zni}D9VT$J"E,$cpK\A x"&zAV 0)![Ja0o?W<.<3&UG/沫1{D.`6}!ʜw2*YB}0xrY\}O />$'ulj2x&M@Q` ;rꚜ]k┇3WUpwg m\[kܶRHz!7DLޓpAFk#./f.rVQ~u6qܠSw~ wkW+fpeD>,54X*<^7=qrhS!n!04j`홚уf_kw/+?>̆V&͒+}Tb8_[nk(BV+~3}~qPEr'EY띄|۰۬2gQF'n򞣫lYzWFumwm}V+RT2o,}g2 y Wd9=$dj9im½Ʃ7?*EaϮ.$weWywQ5kʰ-S1Q9<! ɟ~?}?域?~׏?}J?}G e!79l#aJ/!ٻ6,W}Jid2$_$({IlE-ݬ:un}ďY?[1U57*֦juMnkrZ>BpÈ@6Uc&^.5:gKtc(PYh㜌ހyh&2K?+à z)E|xX?1ӻF׭Yn"먕)/'$W鈐Lں-ŪS^tdE83?&f\Tl 0imXy !xfmDHo[w.:+ls UÂ$Ы3Z܀7Z`5:^m[a\fYv]wӛEFքsc̱b6b;,%"@ F/pJ൚0; fbX饽M24zvn)/ C!ʐF: lp6HFjRz\T§X6}sVHRR,GtzQ|{ K%۸V \|w}Wojl~~ %ɾN8ROw'(}9?vm5T:ShgYj+Ԝ7j[k樮m3.WmDt~^"Oe\%ͭ_I7g'Òe?PK@AF2wWZ> ,9}U'*|+RҵMX ,S2I8N\^q*vZiGIR2(HV*T;VYbw^j>( %`JgVhꪶnu E|7?agO'ݕPSꌒߖν7\p^|O{1|gv?N+}^E>H ׃wufcqt?Y:>7a?= < {EUrJR}pd'H +hv1pR.@ZFԹURBҜ3D/@`+H+ /hˡ׈+S\%0~öpuj?)ř]=vpuc__\+WPu)pSS TXupK+iz1prv1p. 8SJRJWWTH&尫$^eWIZu*Ivp ɧ?1?IU;^f2M_%wOqRΛ\a2\Cv pz3S| ]X-4XݓWฌNX}_7Kޯs^iUoDz 2NH^1O%$g#F*.tϣ'p3[|LHUqǡKQHP!oX<9. 67_*\ }߇{33=oi MtE^\է簐zm` XLG"IH9Rig {5S-GjĉґKs'ä>KúԔ,y0e,^ʒ 7g.]!vп]=xTCXZ-rlĦMoWKƯri;+7go|LHuIV/WSlkNakƊp\=T9Q ̍<2, D *-g&X`T9F$jCXYL^9%b1h#dȘtflK(@-ܔњh=_m_|m=I7U,6^m_=e~oF\ތ-(xȌZQ#ɃeZҜ=R\q P upDHs0t #E0!kXQ`5,9K)-ZhJ(~U^,Rg2DO2"k/%%evXwK 2(aHf`&n`Zo !|9 AS`]s(nsi%v4K.lisC{o/_ᅁG`]L7gքҡ :8@CJ~bX $txFٝG#t!DH-xT Jg@iM!) G"S#=dAD׍yEgqqa4 qM^G* [ÝJ,5R)CDJFu>۪wȻEiR5ۖ  j/R[2_i/=/3#P:Wz|eR^ѧY3$-?޸ H;' 7MJySLoU|Di2FP2a XV !do lT1==g|=7 ~ۋѧ^Gvq;ZnY(L坱0?sK5"eiZӀӱYiر\*3C?F FuW=\ZDMŞ4Q9ަՑW'/[Z4lPՏ ^eVg+wCGʛ%jQLjJ;e⾑zoU8$ §睍?'♗&!]'-Do*R)E"P8;|?k%J|֭&皽[t0EPzXzx.ޠ[Nv޲V98ަSs`4-Q?_+V_v߆rG1'wIclQl#ɶWRrN'ۭKq U_ܖZ2o)hw'>9\vdX};st>Ɠm>._g=];t='jE'ʼn&7k IgwV 3\dPhN\P3.u؅JC_ }qH*!C`c.83XE8hDZ#'O}Nt1eo=,vɧ7jm s?ϖ/N-+D4&FqDRD%N"5O %op4*kc@QAsglrD`pХ ,*ygg ȳu<fŮ.Du*{^iY D,烲P)YZB _¤`!ˠMz sZD)Z0la%^e dkC8𐊎ҴnY~QIIӊf(Q>(y&2 󬄋r 8M(rneZGEdQ0A[x; [Lp9q-~jlֆwgo, Nfc4h},M>D}=%M3}1c_J$<1IJ1_ckhjhƷ0U/[Ƣc\͠c҅Vpb#V)o(oi[饽C7NC"ꞵICs͔IAYMK!zƌƁZcĜEߑj$Qs7LĪ*,W]XJ (V{ K`݇h[6tr}cڦa`d}8sBML65<({ cN:т1byybJ/WSeZw[;4S,Stog d.&  *h ϐ4N(|G:ή\0X1<:P@;sR &^ZLFJ$[CmLBGC4ȣ(‭iU"bK(F[ :ќlZ2_RdM;٧~=>kˏ0bV4qoag[%"$㫯Wb*{|E ע:\?zBPՓH L)^cj<co:gJw|'t GϮ>aǣ>sY:A쳶Ժ6D)@$=b"tZ+ "(m GR[f1yd!ap) P|iO L*#R"ALNF0(Qi+ , L6; ]J? Cua*HA%Rxa4Z+c(C0h @3CښZ#P)h:@1J\4()0+T)LxDAgUiBX.i ٽ*U;Kmݙ^ɓՃ7/.ͽ0Ws{I {߷ E8?ߌ%.\޹IMv?ZkZU _]\]d 1( ~nѣ:WYXY+k5rU+j%Qh)K> C)f;D`xPu_hwqb~; nFbUW4s,xR9$w؉t&T&yN _T3&vǝнEw߿??|8cܜ˻^'q N*xS~FE|QU3QWX߬ja9U:[=gˡ&H37\ B )/kmemb7!lE";@ U0yg `_v*dgd2 Dh[1BmW/(mu^̄6[dɼ*M9",G%a}HѳyfZnߧ#3$$R'$ 'NZ+Ar7V<8-so4 91&Yq! UQ䵵͉k/lO^;/oM]Q j5j#:`{ $ew ] yIN6)(@^Z=būI-"ùzjQ!rV*|26I/DY:uJ.%%37xlJC-8-HgՙY2y=W5;8k癎 6 y]uIK<ÒNiIlV=^zԤ+7FtEO,t}.Lg VQ!<Eϗftb!o53zoW/o쏞(t]7͒;Lg]׽_|ӅOIQ›CI7W9?hg|rpknͧW{oujws U賏lQtC-Xnm~ \RA=H 6)u(u* xLZ0ޚxBľC}'{K AGHFz }ܛч츓2dU"[Kh^tg. ŇzdOY6ӵi1un=wYU},/jGҎXAwpq+^e>OG5q?D͜$.C I$OxME} 8mC6¡Bi0މ,@˜eʤd-*ԈoҲ$2uhXMR!4҈^9Y&29R4-G/ Pb`&QS¡r"uzS^6T[^hpPL If#gO19B- S^F55nˮ.q^|X)C\-k'[.k%ۇ[6ɩZS٣ʉ\+*?M #'QSoU'x,Bh2)z^*)CԊcDʘ5sJ.3Q:>sr&~XJBz4ʉ$g:A*'c -uD׊ب(en(ƞ 64 :mGlrHmjFyL̦vm) -j/DfԓM3&2Q*,:Vc]FvZmbjY*7ZȨ2@+:kbAMP8׊c2ҀFuՌ4VF}avK~)Xe͈([Dlq[vPIMR 6FQ4ϼltm;F,,QHFҀq J3u9DCI%ŴHmGppn"yo/G\u7鬳Y).qQIj $3>SI -AQg W]bS<ԛ ෣pUA .tg!kwV:lr\>v^`OE)KL5 ߼p.y ~ލ>d.~s erfW=?~zǝ;qNJ@Ƹ.*(YS@Y,3}^Onz ?O> c+tN6f|I'ϻ>gM†)ݷuGշTJLgR_+U0;a4XA/?v&7,wnz=l nsQOݒijjb5U e[CN|vʨ'扡HN/#Qdellld3,HmSLfd9*)jD Ny2 .z-m)g4dBS9^ jւK RTi]i$%xv\5[󴥋84)67?`"33C6< [_|o q/;)jS](`I 8dgw8i:dL(vֲxW}:mMV+O`h%-v)$=9qՒ]f=%R,jxѢ" @! 23Z4;B˫YR8zvjcñs^D@5 ) f|L.mI;ɡ!L*PPʹ/f{_z/c8Mc8/t.Yf;{ Ηuhc:'(hy>ϖ{.CcI"h \)JAEo3y8tn }uu mkƄQL(k7a`_eGzhQ KJO@DLЌhqϦRp hlFfƜ1@>;c׷q+R'B2%;IY/ SYrחSܿO6Gs,KԿg?Ocg )ĘKURz(FSgBa ܰ֓U'kunFxg5i1oͰ@N;Kn5tH;D u&uʷC%M16e C-Rn6VhlfgzЄA$(IHzK8rJ`{&hkgά01);E-r=ԬgV@"B l C ~'.&V3◀ 8JP5'6 ]V&ŝ |_#|\LUZwktRVx3a'sPG2\ғd `8'Rx/B>u .whE#uɱ9A&4P(e4{2ogƥdJɸa] M Yi`YcL(yEۛP8Okd `1}֤fFSY S=㴞\NRhU`1:"ca9)ZtRQղ|,'Zk  \q9*o:\)yZfJ="BUvU"EշWVI!wqW]j;#׏~ynlj v"83觍i0D]w(9"K2 f!xx۵ ۅ1-[5F( zi֐iNєfriyT/h_4LQBJSd2l ;uFkmꁛ-1Rڈ,^S-ùpn ,s$@x-\Z|u2[Q:F;|mԒqT' h$VK dfe xԃ @LAYE`,ch|F0{L'(07hLlWy,&6J L5.R њ_ z8F l/ˤpzM;p3c^ .*2iUT W \q˙GW(0WhઈkıJ+XH``p ="BI*r~,pUt*RV ~ ˰*uݹvQU"x3%W*" T1pƲY#)D @Lҁ`*0C^nyV Ύ{kX6q>l.tl-6,U P`Mh EZkW(ẅo4堎 gZ{_6lݙN03 7 YXӲdwwpY-*,[F [U:$/=$Y: d *JNWmY] \4B EWAO,ڄ s+U*th JQɄ*u;6<'0{iZ!zpU-%z(m]t%;ڷ)ܘ փ{luU $t(QMJ0*B:uh!m+DdGW'HWZiBt5M!t( J]`Kd2tpMTm+DyGW'HWe KBTBWV>DRwtutnCh-)¥~QnfJs< .!B&DkY*th91m+DI;uute2`[ ]\Fe*thyЕZ]!DIF]!Z+NWRn$JQ-?V0>Hx.FV΅H: QXZfʹ4Kj٦D]m? 1kB՜k΄BFr4}.è]|lvCdA9)/ɥ*%op1 P(3Z}z(y!B}RXJt•*Bf莮NO5I<_^)#q8ngOg;BE.z7l|2%5"̈́w&,h\iw Fv8o^EOf5c7ʺ8_e9 quf w1SrZfVXF5=Kh}s_Ξ-`ll<2l j*1.vBʶcYۇ*ռˋjnvV+U]f4&œ7e$>x=v.o{}?lv (V߇q 5ѮZu\@}l?R 2VT]-87KE'؉՗ƿ_.{_Vj꯳@LWA! Pv^IUS.K9jbv{t`oq:}Vw>R'XF|8p$pyt= rvnYg&Ԇ~+8=1@ߑx)MG/Nِ԰(d4e1|Lx[*tAkx-xEF7U\+q!-ܪ|_OMOx}koL(.N诏k6BMԊ7vMp(w37W_޺lj ~ߠKSU|9>Vl^6`{l;6[&h+|w/U󉋛_G",Q>I`|\>+w#rwHC&wLjJ?5'1kvJBypQ:Fs1 S AM'~\^26 O3ddPjŽ"-Zu,B_\.hȥ !R8 )Ļ*tښݺp;%~>k- O( [L=YO?!JfLb^ho5JT.ڃ6eWvyƱ=`.%؃j-߶IIv(e@D}SY]ڄH JnJ31SM%i~+bBpe1^@؏CYXg2ւQt-ִ>W"KjT'NP &! f,B]!J;:ADiBtU: D%)ҕZ]`ж0W%CWִ%':E2\HP;V,upE2yVQGWCWkuBteXt45trC:uut @%Xڸn7.S,~僯*'씏?Iv @G~y2Mw5^(/kWKkݠ{̷R>p7d#v`z'eG_]]/|-8cu󊘦uvXѢ;\~ܕ+R&7k!-ik[^U_y]>,V;bZw#fs,w09ɠ=vLK&R |T 2P>@K2>eP/lT ' N;Ju nq^Pum|Ό)o[qǻ]m~\A i]#1:ro7!R}nqK)vWdjlͅkpl KG5Me熭A:3i3$NADC:qIH=`PX[=lahv?2_ xal$L?5":t[iPD4 ƩyO9̗3KNm>C6+RbRn<-(}H~()M{}W{v@  ZĈ8| ] MC$[M$X8?m^=Cp6?ac^?+b ś1tW=דA~wޱ~Gomr7+2ۦ뱠˾\%+N/ 2~E0Z,R2aˬ,y )nJl`fa9q\Hl!dJS&P>X Qahԍ*E.%eKu1^ Q(!d`$PO1&`#/Mh J:{Пɤ >[BR =/X&ݨFiNbIb 7s{`-0w`=/UP5VnMo%1G}O|r,e2_,]{3lbi5LgV a1L.<]v- ,b 4%py':"N"xά,h*Iwyc:'O=A۞M5p?gvvLnaR 2}J/ 2FHk&sad2SS;l﬊ReAȅ3 "-|0":5c2(S{pcpugЏb˄dW7g"#qEY)ԖT͛9[ 4s8`(&s-Ve G|g)]̗C͗Cϗ͗PTZDR#4Xy R,\[ J))F(YMZKтgS]gefSa gϑĺFw0 2]ZoQ6OʌB5$(t4Rd"ymV0|ٻƍ$Ug7Cv 6 2 8d`YI{~$%K(eNfwaeP=: -mO}Ew:H#_2.PHx3jg>sWL#,8ZI7)&x%\S)`wqf^ كO*lC_Jy(> E6vi=2,VA.eܻnhzջ.FwECppIϸM79^ e>W?ϒ5t=_މKju-;L{2P#t99&U3: WpD"^М1E{qBI{0;J(NJׁڙZxm5b0V 'LԞHn#fҨN~:@ATl=N;[B7zf!%  鐸c_g(Q/`uv3g- ̏u`7[=zhc>a*yž^%$}~ ;]QH΃!g\rt\8L A/Zja>t)6{5"1,cCydX*AU[&ѯL|} v{Yu ?o?J)B] m=LdZQ+ɃeZҜ=R @8POL}ze^0&"E0!kXQ`A%sa[e-q4`%(zᣂg&pޞ`'$d>$Z]&r GIv@dj^wJB=XxS2kBSPQH`qxIxkmέ† gH&+߃FK/MHԂG%Zbr16CR$U,2%`QKʚJz{UI4/\zHpxa4q&#PP% Wc!"#U٪0aB5tԓ@Qךfp^=.xZu$B>M//ݐ- d8Xyb2:ldBЈbVqCTJI/Ɵ.)f&>q=^G" k5)EtPm9T#N\Epˢk-'խ=s>_s$-mSs72ɏI3V|P&LsYa"Σq>9l-VcNJ2fO!IÐIP?eg8#-.e;>< >3~&VZvÔ#?-q̌hdF3m90+̡C]KFrHʃ:ʲ}X2re ?FF(j}W.=&f6!A}6}(ܮT`j[^ֈLnVj?M?mF-'~;o>x>Nm/|ʌ5 c6TD7FlAtWRrJ]˔<-=7"Umb0UT[-ՖĥֻCbj=|l՛%W r;(r?l齨\oM1.jiI"YF~KGsKlaIs[+61хa7kx_rRV=߾q{?iѐ Q) FF*4jF{#zz's_gG^xnc`~π`JIXy' l|ѳZjID!N. л|u# t2@י%bE-˱KoeJ'T!.  :”xBBRcȳzg촌O_O|ڥË尵OGjz՗UwQfL>]Mzn~hx-+-l m,; ;!qTINA#G[TrX`5 ɹ]J1@|iNS5:/)I H):&(fEp$VD0B ح"sO]JB8~XLb6Z+{aT@/3gsmMa4H@\ E %.GqRK[X* b]HA;{+FPyk3(ڒo]_.Y\9CZPT:?Y|m.E2̲]*1 1LB\cm]F1쫷3MWnF7Rڞ"?˫b[g")[S&:!cgX2S`)KxMSwfpcg0y}m[ `ۊ.:mqJ|!fv%xj[\^YYŧ|?2f&+]i;ؔ%wO5R}*gHc@)әk hjB&[+` (U&?e+TΪKsOՕ兯ogWcfꁙyP[vPSvVVQyeC!Г'7t i7wB2 Fr5~T =}sp3m{8{%h{lY5'"g|9L&.Y3DBd)p3tv|??%*mBTiC2g~~KVE+:D&隸Rdˣ T8(˙`?^|ߦw?ۏ}װ ,Tג)E@ ׄGAe-mFɞ>|*_/ùa 52`k U cJ#68R$#Z{1)վ%Xeȧ>)rp_]?_!u;aC6%|곻Pd3hHpJ,VA.ܻ4R\AJoϾAaÅy˽n#{ ;Y蒋 9V/A۰V%dlBJT./,5"Z(hkc2MLŘqԗ6N]Vu\=QN}k t߆+]˛m^i)Dƫ!\>~N)đ\y$s˕!{eq:N#^{QMD㵷"sĎ'/ݗGEqzw<9!DH"8BNFsA Jn)6k x$ )[XsZ D佖豉h W M$GNe[SsmSM۽j;.z,dtH/k>)|+[ks~9£Jc1F1s&qx )&~IȒjOֽ~j,h3JFHi!Vy歋]iA*j+OvOҗY%zc#cu|l2p6Fw;C;O-nȢqsˠ#<:Ʌծl\%OϗcօpaIM k3NӌYh4 h8`TH2c DzJ4ApO- @D0j@{H@}JRω8ؿaBTBs-uر(QiC>H xd†HVcHjvE|0x )2׊aaR"bgBja~*wC:''R;C~`I Eࠆg93eːFv{5&w&ob}y ,us/W0#KT;c׹Z<ee^ ʋʋU|tTtc UT]g0TDh6]ʋ_WcCy~ [1TMp+~$6tF2g6sFe+g$OHJAH~g${bS2Wo]Z' (}4JW3[zLa3t5r ]MBNWe7PFb& "]ߎhSyNk }:^ʑ!'K}',|{'G1dc`/Z%d۟?p]'ZjE)]ϟ ㎮:,a]H>`ܽ1, 8 O}c˝L y[:ȹKc*gGqĚwG!6Wl(7⃭Gنj'S1]o8\ÿG}u}C{r~sW?N:9kVc%qnSRfD?Km6p͖3dL׳>?"ɕ\MKE,͠9v,jnx`4Ĵs?G̝Kmy\$[s`?4F66E%Eg: ;|X)[H)=38:Ck}ƄjkN pDKQ;< %BRx=8u4h4ZdkLm3Än$'c#/-Ukrs`"EȄH #k-ġ;T{|H.؁03Y/ F*UK`#SH\,IAo1x-&{4:`N#4veۦjD(QTzc<*J|XZc k<V]`&S1USԚfBa@1EjA2\YS r6Й|6@@0JoޛZ2_ ZM=n,\ JAhՑ]p;zu J*Ho{maFAXyǒTq !1T(E(X(! e 5P #M.m%z|p&ػ &/ eP?{thoG.M3i1М<4tU)ugxrF+8}.vv0]=x1n͜񂷪`*V=VuG m3bka%a牎bPT8xi"m6J96]e i<1YE.!hd5Bu AB:Btm(ߺ,\d!:T?3FjM''ml|Y"1J_C6)>^/.On2K,Ƒ\`y6C@FD6Q=J.zXQ.x З&28H Hy1 j(] mM3-Ah_1S;6Ej-)+Wijw ,M90f^(Z+a:#,$yS:7e6*I@̴LT3oPSd~A5qw pD5xޜ,ʰ&ǂrݢ aB`v2Ԛ*G4kUG٧F; j3y o;TJѰ<5{%E`!-jw;$a3>Y` pp_řvv jf!nlm9ey!<-nɮ~^[ әdHCPgH7|lѳϼ65eogS bq15dݨ<[ 2Ęx1˳ mWidN/_ /{ s˩4ts ]a7Vho:gh.Jt;cNvp l )aTХ'6:Rz dP >??}šKۜRawdUoFi5‘rO(c$T:1|4#Ґ]ORuM` AqMic0 <4KbCJm>dxUTAִQkg&]l&1桭 j>ڗ,vydރ P|ƍA-,pQϡ@ /5V]`"0h| #+υiq0]p(dW(8q˹ !nIr[15 i V8FlnT*Kc<Ttb4&ٕZCp$2ubՌHܧ(@GQA%,0H5'[F &|f4AyWwX0Og '7෨vp^ru͹7ga3b37~MnK9 8n 7mX/] +$&^c( [MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&^obhK0`f@;&/B'&ʨzI ܳ&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@7 4os-%\:N%$C'\B$kLǁV@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 Iכ l) In$A/?$P{4 @&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@' ׭֋7|6Ϋo7pݵvqM=xv`=-$>%&\ț .AY{%tIK!G G 8n'g93 ]M.;]MSz=t[>^yf?o.Y+kTg(|}R{o2 wLэ`ǵ_F|kw5fȅhTkxK.7]X;o+NK ^Ec^3e.S,[ q=`vBr+`է-*s.??7S.~]ӧ/83xsD,kxMD([HOӿNs[;& n!W29,?+ogv>j4{_vSإnhfV엝%ڽԄm-뵻yvIv4sO3QIFt_2ɏQ9%avgv۶dBrۊf+o5A|8(f"Ѣ~d)!]PWd2 IӾC/S9*ZFX+bAAAƢـ!罁F&!;+W eg6szʦGbW~RrW&R6=WʦGYozd(ʃɷA 즞27L,ڨEd堗5[HuI[A+`qM+u>w]1enيv+=^+~iqS?8>6uѩܦ]DW A 27RtŴ+:* ]0+5bf+D_t5C]^yAb`q+urp|t\@ tыv(EbZ̾()]sԕ`dwжtŸJӆ'20eGF k-FWki}]WDJW,u}0ARtEjg\/i#+2M+ҕyb՛}krM(`yjnx(Zo&JQbf[J2EWϭz,btŸKӢ]WLM uC@銀1bܩ7EszW]QW&j+^NF(ZmbJ늮f+GIcq]1-uŔDWsF btŸJq=4Xt5C]y銀wŸr+u:w]1/uN؞+9}W^+Ϳ)MBA9׊>A %*m0rbdKeTa32K/l_u7'{}%f8]=D:btŸNI2ƢJ[ ]1wbtŸQ}R]QWtE]17RtEFuŔf+kG-HW@N2ȸic]WDf+DТx3/+uZuŔDWsԕÝ,p=]pPrFČ 2q)m黚<:$]0JW2v+rA Q֩]1%yWsU|PVI +: DiK]IWGqQу 9l4,iTR'nh.UeC 1ҌflU2,Q*9FUr͊D=1pʭkw{zqa{mT(!E݈ەU btŸHjrQܞ,ډWiA"N׈Ӿ^/Sz,XAb('"\1 Z} ues {qALtŴ:+Pt5C]aPVIq]+w%]QW.6 2n)"ZT6w]1VEW3ԕAR2cѕ_?#EWL1w]1e(樫Nvub6f?2Hz0Zt5]ŨCA"`P^wtŴ&+e"ף+Ī*͓jp:uF"L8Jٲ<~|s"*t1ႊRtŴSMdGiU u6rbtŸi]WL%R(HW]1>dJ㋮f+T ]r+ƵVe+ EWsԕS~^+2J׊bک]j@5/(q+:w]1e(Ib`/q]1m~d(* 1btŸi]Sz[t5G]97RW|6{w*AV* lD⟸D< iY"> ;4Y?rh x]7M+&kuv[|56u`m`N=w+1(FUA()pטQlUdیWiO<h\u*]x}VZ3%be֋t^'nU{tso[ʋV5}G~~[1u[' V6h]EkueVhƴ{u3e=I)M9UmxuЯ8Wݺy溼?)>9#u>Co}ãO${~[xc8c|$NVׯ^|sd>Ki_??_F~\]SbV,mCzҰ 3kjRtTCG~{8ڂuMMG]N)$QEqxqptBwTu`ўUv_UZς5NH@zӍ[Zh ^7i^0>̠pz}}%z<=C~W,_Z^]t^nZG/NO~wu}_a{/XnϳyQ,-˽"ߪV?6>? ej)vXe:~Qß\ꝎjE/^4{MIϷone<:>>{?*듫`poo~ý됤ON8޺ 7aQĺ*wQ {?RЉV%WFo?9^\NóQ{m:_-]]{×gT=*u1.F%\t7Vg~~+*m//1#̞m+`j|ITA@iumrK|n><+y`m? ТPTSs5Qtb[;hn[m5cߧFXmo\u֧^uq?(mFhZ׽j76u}mnam 9|Vz2jOVW=1w-?p[7_xcV+ wFXuhn|RIQ]w+u$Eץ8`<쇇<~l+{0^5Ƿ%f~߄}<-7O{qF’ѼRwYi?QΆp:ڻ8]ߴ/ý_7_Jk:o궢x \nhH7#Ϩۗ㎈6?7D阮ww~- ·;zgEmisTuN4wMюXWS]Ι%KQYSMКuMlmo6y z- \2MmIaKa0@~YQ^+b{ m)62fp`xP(=mQ^\Vt78NTy;Oقm+pa }0 ֙2NUh=Ti(PkХLi,S3\ u==@o]SCKwqGOm };@:fjlEٻJ7&}r),`F4zl"D蒥6:(: ;S1QmCTVo%uTlw1ǻǹw3[;?ұv?gYw=-o)]muYk]A[-PGtcvFEoP8?$&n:c^cRbڪi֦ڇKR9'u{Ťt.I3Ya 96@Fu!&=RMuulb 56m6ԎJyکħtA(9\pqsp=rTtPEUvG{¼hRM9 ZC/QjRԭje|6DzE\[iuq+I?>o]Hsƺ<>&>. \_ԇ$_f^74l)CQwI&J0UN$9x:xI[WV?񴩵yWaLe:|ݺUʸ`,゙ XHb2u>>4$;5Qu;:l5\}#6`Html!HY v?vHq &D*-[y¬+/にg7|~d܀nѾ^3e童4i=nᔛ īūkw5m]G+w U0U ɷUlcBt-AQ֙z֓t>gwl?H7P+ZhxO6iB;X |p^I5+(}2rz*|qr<>gIlus)ۻ虀.n&`QC4%3} ]jCNoI$tDeumNJv'7~_"]c(%EGMhݚޥu^zۙ"|tMN>s %Efp,|;Os="ziw,k{~Smg|KWٻ6$UgÎ8ٸd7XCD"|_ ") 9[-OK6]SU]U]+#3^hZsUpAzPLg<_hU=ynE"<*a5K &)")  dJQKTJz{͓V00RnSO9z(jaMm: H%2\}k3T ÕX9eHȻ^Iu`a4-ZdpH>dv?Mnw"3K9xe]rr_=}`^] AqwB_R~\xqaF~X`>¨?IU-JAjf1ZOp trI1ѷuOXe}+m] ~.I[vױQܭ<4)jsai]v[w52D'yG&>^q 0w+kzm}˻p_U]WmK/W~qK;%i햭ӫ ϐ X*\T_17Hڝ X`4+#uhO}MǛYgmuI^5ijf?ǓxDa}wqSsa+3å_nc۔1›$w$,-rޡh67ӆ&Kmpl$ѠQ lW8{:6.{L[f5Hn\qV]ܙwX֜DDFoB%{ :L "`j$RkcXzbX }%K}vdBe e:2p (F !!DEn(z"FpG`RDLRt~(}@8rƽr6/:ssFuW'4Ӛ?˺U F9 rm"ԜYJ-/g\:<7݃`?G C`c.8-;f860# E8xDZ#'o E@Ztog" ;m+Y, a_~[ QZ)Wt<1ʝWkM$"*;>%R,SLy%ịUh4By`eF͝QFGR2 T_Lf F~ zr<W+_==bwg4%Fx\8D@[ƨhPT%h`TjbFyi_VUx*p܋oP18:^[L H !d Pӣ]7.~y{GiS\uŹ#1œ;#Z .Vш- ";9J1Y{&GM{9;ET*J^l# GI {by̠h$00rIKD$2 RjDs$rŐ*  x(g(g\YQӥG."yhRJ"l/tkːs\I0!e."S<&gRQwOk-wSa|xv.n]mp*U*EǸ.'qy10 . %sH[]F-bePv~y{G.tbͭu|COgCΒ\..PHx3jg>LLj9+r9E[=ǻobSy[zuc~{o6haT̡]?Zha+ci4Џ#6w] #_볇=x|j= *RKlGLì)Dʫ\ke4\ZHT3EyWKDŽ Oc*]e) 8 h[/5%<}%9C +>?E6;ݰtiU bKNAh͵5qN%:F!Xrn_XL| ׸P$Ȇu;"E hIE$s2JaQi+!͒$@2?(%l)$6GXU O9†QESa5=? ϮFښ:#P)h`up<'0bh!$ YO;2`R6x# \7˟yvPcϵ "'H lv1e{I>7B)"rs2b$wqKS$ q)3K%MM]=E?_yLȣ},|bv8)8FBWtrbNl.d6%xy\>k~+l[E-6tw \:$U\%^`AGQ_=֤S-6`o4Vi4zk~`G7;]]lԥ$c̱r> c<,%2@JF/tJk5a7ǭ Dqɓ{1p4a#A2sVk/;^U~ 'p<ؽ3?~6~|VڛTOdR)JST a^~gV0#*)dFRKaLvԳBdշz꭬#vѵet ӯZl{K,- zxtJlu(XQ+* E*CA0b[[X[cGڷLzvP(m# X噷.b8wQHl Q^Eme^03s|7ϟML fxdh} 죣ϗPws쭣X94ޕͰک.g N9#) rMKIp`Yho}@?$9Br]]fȇMda^X[YBRy$k(lڔ=ds=5=UO?S/E"VIIج ]l0#/s]$v~8٧D!Ndcϧ !%8"r*X[ 䨒LXlW r!R)l"DM,ƵPX5Վ A>MNߵ%VjXjD¥֞]+qvhvze~In'{te%tRvH^Ykʏ's$[tE(BȎ(X6Z#˕S"5EF,L18SSrfr(,! 4*H1Tkdl&؎4f/bc,T# ^^WfaVwH]LJ'uB./g./o &~d)JGv$A0hEJgZGֱm;`y2#(ɮq2 f >[bN8O7 d]vyloʹTPG6W$zQVk&kKYyLLE"ll`IRtZ"{ %T5dKc5ay̤HTBcpUUUֹS+;+ܪqrplઊs*pR:9WNo/}zir拪|TVe8{? gm{K_$M eW.&:Wl|EkrO T^|P5ڲ?}Z}Ooݫ%%ݽїNq lfuʀn+GdY yQFDSҾL %xZ5ڿ/~ܮt^TߝM߯1ll [p- l 깋ڌi:TC}ϰ)3Js^N: Sc\,j{j(J+ blhlƠr "_.Hek/ ҩ"]&  ]i3qvLmǐ:^_N)uϜ}\^ā ǫo]N/7J_ڭ/妦gda,q1DuS;B2ee(F."W%y-JM_A(#|yG1=#|kȃzəXǤ}:҄XCpc*a#s OEș{lyD&SzHUޫz?Bg+#dzz)țk&{]t ʙ]T!DBzifVic}4cK-lvWт8T2PO"E=|b'و/j>tYzW8z1a|hf YX);,& R Q$Q`tB]~My AYNO4_BO?O~*>}*h}!(}m{z;E }>K Ԅ&X [!./{]w748JȨ s;*w5.U5=S4o vqS+5 Ox3,?qL/gWWuk~OIVwJ>H÷:[Qt\zX3󂵺 - =wӋ{S+s,>YKw;uDkh|tȘw9ޝnj~~mj,Һ̦z\)p3w3uY\ĭU2|r`?ēu- b`wGd=s.=i|-5h̋(@=x~)I*/Lt$fj% 0b/ahA8HE4zw<1Ɗf/(IFHVJ&護`$Kٗ]|Ыm(H7'ϧ-/HRѸے;QۤRVƜ4'N.'YPYug-LZN4X0Af%IQCRd1y;%Xnם;C3tg_S|׸˦_{.VSle'QĞ);ރ&*,k'X)d-j$ ZpRSt1xSjg SpS,2Ki-f_*Q1]L-A |6[OǞ}_9~اXU+F+yʣ=K^}s]s*9xϘL:9ա 11:VQ|'V _I*՞ MROvrB$P!$%^՝z"E8c 9)+f?-ӆ落\mܠ[0_MY(K&hqEcq ؠ s65F *'y"V =Fx&{yvx} |#VA/iWyjfT!kPRuf25ڥ_/.7AJJ4)7ڢ5)Q0&8*$ 54Zn3= I,x$);OtA"taL$ Zd@AAYP`E:@\ eR#W<%Š<7fhg] &:hք/`K;4"Ms*U`7&}D68%$P SM ?% ;W|t)HJT\H9EgSD+D L@FՌj'~ҍv(;iYQ+EQdd-%9fЦT$%l.-'E0G8f:WݝÃFo;?΀wYy23B Wfuf)\T 1/T7lK ьnGef)^Lv+dckŋ4'ir2ZۖE:@ٲ" 訂 d AE =G;!c]]}<h!Z47AuaM:?&Sn^DƟl`lqy/\\-k֢ c.lDR#ת{gAqWm.d߯zXr,Y)t$EVb}X!޵ٓs5ϻȿYVɻ-eEIeEBlYL+<oQx 0Yοf%sK`N19{^2>)8r_Ҷc.p+$OP&2cvq2|%n0x"_5~'6ЁS{RwU$Q}ߢVxq}:xP?{E].ۿѤtBգMK+kS9`6iq`~{9fuίn/|wy~usa@({-̙}8ΣzlhhK^pғbkOnl[7b{7j{7 -ĄIh5}GWˁ]9<^z8[{lwˇguJ\YhrMN4}9#D?qkƕi߿X(^-33rK,~idz6qmڵHeŎG4Sk "|së_P~?ק {/N߼?{Xbʶ+¯@3>_\[v--ӵԋߣ_5yKѯvORB,3 Pq_RpcH"!olŜj&9=6Z4F_~ ns1Cd?n #TbJzX??N&!sN킺|*ժAD!LJdQ1m}:Ԅ]qtyJ/J$dkhb+朇ܛ"IBBTYwB )һjs+!+xju [=Ns RAпd@ڨ M5\LGBKFB$K3 h O:2A+u#уjⶄX^Jfn#\&[{4&Bq5{gNfM/v_mnBOrhB&zls_ MV?s\ltwCw_nwns<=we9mtgCw6ٺo-ݹ{>L'-=~}dhfdzyc;S[M/o7ffc2gS3g R_Gn=6$l.ՠn~(ڀQ]Ԉ(+ȍ6L5d&hP`E * %]J0TK6zkr(")W*fm2;9(lSx\WR0ZmikzR/8kh=zUss쭣GFx.|~939ԈFhPЀBdJ`t0[EߨCvΡ:pphl2K0*gNfaY$c JEMQmY$NȺޡQۡEewɄݑ;OOd!m)i7[HAF2),2%0g.Yl8QhpH̲姄^lXEIR[:a6FvĹ{ikГYӛ^.[V'JܴJn%j=◻gTMZn>IA,|% d pRd3gɼ! Ai@J1fGjFmMN)J!JmmևI<\HTzLJf&ݚRMVՅPY^Tnĺ\\vtϮfqqxohh44̑5v^jG*h*'iy*;22ȨTP@POEm4dK&ce1hle]M5v֮6;ڽkK9-u4RhJ$Duaq,qI-9u-ج1XŁ2YEGtMB@a& _+ǔUb@5+jܭ[P_=Rl46,HgK Xe3ԃĽRq7ڻzgj#n!6] sR+g' !}3 5(@d>z4*4Ul 8YV2Kw v%'`Ḥ .B P(>,R#ZT.R\H|c" B$ΐ5E ?\ۛPM-E͆J(6kJOAd=OpOzћdbj#*vla9):TXNHZYYNB,' ˉ51U)SWu]])v^]};ŭxB&dUW?uUꊤL[TWܪ>=N*ϔjEpid*ahrme+$$#nOv-DpqGpmYH/*Bl#T| D T9k7GGc2AK A*  "$"@]C_3bHKO<0 FbF/3Vp(ju|IdAiJZ7ۖ'@=lNIe9Y[)AID >tt &^~xߍw9q1 cF"L̼2" 1JDcRozˢhys[wʜ7zv(|ckBkwf3di%2F*&=wFgɤ=P"{=`<֝QȾ:5A/f` n^U)CĿV@"!4 }AvGg{D(J?Ze*<J7ui)Fc9 %jZLvCqėyV? >Lz2'+S1Z0-"hGh *(JcTl8)N":ץ!tO+m}ry>}{h9İGOWl%C>۠}<$f5`6AS wK#)i˓xN:@C20T@l饷eB]ӣGco)ۄO@UչyP "2D'RȣRFy:%*EKZy!6Sӊrt^_P\qa(/LPb۸ȕ[4ڶޛ_ǧ*N)~!rjO~&=AlNX?rM4ry'AMcKaI\XUv8{LPKkEP(O\y:?Mt*eTBSsɗQ =y5A!h;@,hn=eLE%>Qgcc*S>RlJ3.> 'dPXI[::kdo7F?uv'j8;햾'O7iܧ{Tp:+\bzLxU*S(hv|&%(Ҁ Tr{V~p!ʵZY#>KZA:=h=$HLid2[i55KlbX.K+d2g1&=:KN$O.2ݟ"x̒s=%*'xnOkݚ]ono[~ꗶ8M^nŲ|2*>I+=ns6gq \qB÷*[أjؖ\ -/nv5Y\n۔|s:$ -Tgw޴!6뱋<0eM*ZU ݘwc>g5YEVYGj@޴ӭ'хŠkY\sOݨ@ͤc`j drK7"K@QDWIn ԳQ_'NGQ\+D) $TPMj@ 6hFe NH@$$J"hQ2' vڞꈐ+'4##" Hbj,neB6Jg!b>X5b88s07&!e( HL[9ՠh,Kp Xq6y>LGfL.吩Pt]Шqd ,XXm#[tJܢWsiF\ ψ㼘5f˟!LQ@pN U-2U}>^$\!~#a' 2Oab 2O?}f[;i=ltC#_T3S*:HWT_,zSlЫ*~rJ6Mh˵ͲE^._|vJ6,{)^շV&M_Xnh6yZٱ;xV 78@@oT${]/·y}T 5Wr?\ W!DvnkB0}gYSC^ Ua') ne)DdOsmKxFպƇgs~hZT/I:u 0!"GV>b^G"3EM}DD<2 \VwV˨T`j,#w((4Ljj5PHB £m<TܙQ[BDu⮦nI(UmpZeGw],lAKH4QKȌZiBJ%MDBԑT<]!Ipat1uj_z*0sJAhô:iCX7Rr -ݾ"1cEP96F}̺ *b$Dgq7VCM17UdJoU͛B5JA,<ۏNpȦMƹe,_m bGLTE9B*ViѺ[2\?P=xw1,>xݞ`\O>dP-k{P6b^y|?p`V2u$!7t6 kFsX>"#FO*+X|8]9<η\]o㨌lyɶQ[Jv-rf^K$&ӐJE{Mh1/n7v;`0,%N-矊/~7 g!Q\5^`)KB=8Jk4.6I<+öBJ~~\O|6_ݷ)3wx e!wm=w ??d9[Cw񚡹 m|ł0.]r˸߭+f9n81!e^X1Hhc IJ YUfa8J4`n/q؟ q  VFb A[Vc?L&OJ_RKi5JKd!)8&3S+uOmi$aYh!9pIr"t!m. PՊ'%։!̣˰_By uPa^a\w!WQ0ou]{'J uSfSНqadY``R %OG?,21`$VKQcZ˄}ObxJv-N')NQw{ uˀz1"P9c2u4qZ0WFEL1FT9{)`Up(VC1d9˧%8*o"Wr~m}7OǓq |sdx6 G˼M˽pUqD3/GX%A݀bF@ h]5~Tb<B~]*٥=?NI\&]Cil2Yl=bY4+!iKr}dK@D0$!:yPJŨcAzλb#B_6feVQa4eQš aR:" Hm/rdFIR8P$Zd%2މY:4vh,m2!$h.=Y;%x)#=L y "'Yz' :`6'J#"l%x-$Ԣ`$-pT 0fvMzɠ6,MvTlHOZRnIUƝ9?=DzPyu>IA,V J8ff0yC Ai@J1fG?"&C`Vyj&jY 1%L.@&͵+@idlL؜6Bh e…X˫b/)k<:{~x=Uǻ;tAκ/X?Q@H kRV*(/UvdeQf+ӑ=!) l5dK&ce1hlÈݘ8;{ibn j7ۂТvkާRfNCaD͸Zeg!IQx]8Cme8z$֖f-ج1XŁ2YEG&!ZQ%_+ǔUbH5tY5/=݄ӏ"jU-"떠sv0%u"Y! zVrd"cDrrƴƨ@r2F8RPF%^p!6H1qV#SWmZlLKa\-.x{*C@ 4C$( T܆ ddWl#KQBQDйMacڱ-xh7r_\LяZ~:`W{Xph +qi@ȠV_ƙR]`e&2-NVw~gWq2gt@Ɣ ;,c]~J19a V:B3b,{A dI0XSI VT) $Yf1 \q5)hx7&A<_V6 FҔ1gSŋVd0}\;=t۬ERxK]7sSƘ1Kop&hz) ֲyh+ٴUY*вn->R@6H`ʲ|U>[,hz4*TU Q9QYV2mC )xa1YG(]R.쟊NǃK[hQfw*ŅG<&BN Y(X;km,[%sYMl BbJ=y?UrI "A|'%(meÈˍW_x)w>=4`|bV 2m+@/ 4goTj6fW<*0fl8'uIE]$ukQ9`#GF/İj0k34}O]|ԶP|S=]ld&6+0띑gIi&1 ,yFFzBHi%<]W +$,X eo[~LKԦ $LdŸ ޹v4h2K2U6@F!OU(Ch6|J4v,iOt>u/O;Wz=:7"(/q{XVǶ5pCOti\*_щɕ8Ǻt YÇf az Ad$'JED*r"K E9`R NEoSz|7:oml ږ<ZCA$3AE\w%hQX"6h+ Z %vOMU-"i%WEJ.Z \*+pUuvኤu•`vR^}XM ?{tz4ޏ?)yx,wē[K! ;@v"R]e;_E1"Rb3KN$ JU*w+eEVAp1C e \Pǁ EzT:$8$KM׏iL5Rz;(tcESkUj%#^]6ey.5^u*,n>wz˵蘳8Y県nLFIY(J(E?LaFр2˓IB:f^GӒIY( ɔ#+ D![$5dӈ@O m[!ZHqk_`R2u*;YƀC"VmjI% z.E=H++ٵ UQB|[miyD{iDCV,)FHBh $U Ff&Hr;Hm;\H4)ZF' ;05~~snŅ}e Wݳ4ZjF'6W'ˊ ^vjf !I</,\2>⷇e fmw{erve[⁑w' %pXB-W\ҭ. &e!%>j|n i1GaHe"3רQZ-n:wr\{T~;?=M:^w.| $^=|߯FÃC0SB~\*jrrMJ'$>ⴔgLs,<I_Fg'^^]7n Bٙ0n/uONG}+ݬoHz?^]ͯN}oڹg;ƖG(wΗ5#773,h &k,Xf~zY}`i!Y3mҰ"[P_ r}鏂It쭛"N}[)w~e/c|{=[(W薩(q/u+|;q|еHe=IS1znot4H߼zM~osa_/i. K4]/a?T_#fjW?4M5MMi5_]M>vv](on8/e"׊x2O%ٖc]bS;=;ݪ.,~ dKDeNu^$P2ʢ(+a ueL͡0*U765@!{ R=~Df4OkϣO4]s:rԣ8>7Żr2l/&F .ilb.]U%'6pԓtwUk@Yډx<{e`; !-8JJzOnU8%Rj,\٥·B] &{dC*1JZQT\$5f,z%ÿRQX%| Z (YS D"R@RspEe,f:Tg3s\_ RA׳\܆ԓE g-uѿpWomCi{\NMͿίsWݙ_0QxKOkSH6rii~Ŷmt꥟g3'6uǭE9](/p/-;n-un{\/ja6nԳ{秓_F۽^jH?*tAGw%5Wjo6߆2Kש[osL|fRLn~"nn` ?>a4Θ-2JJ0ءސR }c{:.}!)>CTƤb] V Ͼtm NlE]VzZG.RpSJ<rwnpb8#/\ļ㸧 Qy+hbO2ՑuAd& =tAß݂F8p Q Δ"&YZaȚ-ZFD9gIc)ÂF ! 3I zf^/yUu X'U2"e}[fUep؎"JHdLQmTE)V{MFO]+svwT8=R=^;Y-Q˷[ӢjE_}U~FB &3>N@z4) ڊTpD!k (\9zWr.›_3C̔qT9UVZ4 ֑9#c;,6Bh   r]o*2I{.Ͽr釿hxz;_l4̚d#A&MuFRI т9Z!4Ql(P^JhDU L v‡ [o;btPsͼ\6 Q{@<^"vʚPTUE'8"r c]LfȌ %0NǚL"YE'^+\LI5f쎇=ꞳL50KDlύqD4CD"e ڥdC)i(FC ^r픊 db  ^ a07&ZbI431y KV:M#DNG'cbj-$̜FU^9u6E8.!.qqGc`!%EFz,9Ϙgi_O$@* q8D}HcOem3!=?ԗؽjT}lXFi{3Z\_@ՋZ-0s}3'bc͢Jpqǵ77Nofw^ͦ_.j%/kzM֎?6m?Ø:ߙ.Jם-!zd%$2֕J'" %_CzX+wڄgCf(d :$tr' G#U$INu΢- Q6.ͮ@ւ Sfv)fˡ+R"(^Y$|EƤ9{X?o vOIf{asDAS.!G]H1B;1@jM:Yi,wW\f/IN3GdX%<_H-8ŇHHLIx:(F#4{{fL̲WYD210 jZ :") * Ab(}$d\dcĤC`̓e-P,IJ$2xST叺֙ɭٍ9BwkTE;eyc*vz6ѾIv$˷c'q8O/?~&ya}uZJE| uU2(iJ&j!AVŇLNd 3:J$I )gpTskŴ6;q`j}TIAa53g78`5a{[mSmeãHy=mƟs@iJ8kAi s>,ieDECJj`_:oλDoo-WnܰajC08]|DFq$>XyջYI. |Ap@ Yc@f4Jtd `āh:`I|z*]ihIJ5ڟ骸,YySWT[Ҋ8]&|Jkُ-!l9xD 6JW(P%(T&I&Ka(@ǂ d3}nOGWgA_jal(̉( h'Cs ÄAM;_Gt]:!;~!uvvۀԠ`uN HONvla\ 9BCYK>-F{Sab)AQ"8 YlClx#Wxlp~4X=yk!. E+YR"{0r. ʪU-Ma KLQ[j sJjXR^+YMVPf^:35. -֍>l_8_\:֙v!p S],=cݏ D:wBF …J0+D/Pu04O& à A 1eI JQL9%9;!ɻEKCw@2 *-h=iMc"p`S%>:-RkfQΉc9ix|(qI5i4{t 7Ir`u%_.Ggګxxw}`U 2'r.ca/BG1D+P67Ӄ~!T\>h]aP!drhXEP)=2ٶEQvJ+dG&顖dp_@ɹ MIAHZxnϚg?o:Q1>] 6_{.noKzJN )FhH4t]9?z8}b/=?h]?ED9C Ku{>LKd2@fP+]4{ނo e!DUBAEC4̸K$.S E枓L8&ߚ|>4?L,0}HIQ H2bTJĠ4݌v3>J\九O6|+}Xp}SD2>a$Ӷ,0cɎV:1)PIeC ݣ]c91k!qP?$ͼ)7&]ߤ[Uz=(!D&.<]q뚱5☟ :O̧i(ٙXW:kJaDՉ¿&/QRVGS{ʙƒ4$L3^az&ʇb#Ս }ZCT´s?F(tN9Ϭox7gϼ]ph7? OfM+Ap>'> l&XͅWSMZkWd+C5:{0w;i%νhYJ~l@L!&ʎQp Сe.@^B\Qo9:L8`63ɪ6)GC6 $!C&ӥHEgAHktI"}DomWXz[;dcLK?_|lfzf14z}8b~=(gbwy}YhQK[Lp\.p.؊%CD Bk 0flE(ߛBHI%%kN$%Ms5y(!W/csW\z؁;v< >ܺ-ȕzbQ.gU]sUC͡q,^j &h>;/Oˑ| By5ᦝ6q)I41吪D+9888Izqy:pݭ#}/2r<{'ՂlmdӮf~6G#.ZI&d5!dR%~4YX?:~J-|P1qdhP"EcgvӴ_@!GKcxwMwֺ/8RbZ2NbB$R1X&cȸ#z(9 V )}Z|Lͩs+Nb]MVݭC18/4ٲz{uخobr(A,v1/R*,v>dx}]k%oOe- CCCsАoֈ~#~~M/믽 6moqt~6o=y=LE_hp~uy^<gj&chlͭaHEZ1uF&?W఼|=؉n4}=*y3og~RE*j0^qOo^V/ٿ"$fxw>P[2VV3?;.J]?L/nd} R9`ܷ hY^V h7`Aߞ~zvdn~Z(-_Ǔ \oz,^Uyu߫~6  n%u=8;)n-We!Ud{)ps}H6=@n^wy=gϋ}gu-BY^;;ZݾT9/{LϦW_X'ЃmĸcD%#fW{/k~UW7~al; (hu~t+3)9j_WK3} ,~ꁋ!m;.o-v!X>?t~_<˛e0Ֆu xW2HJqjՑ/:VdzJ8_^=PUT1Bw4w>yz0n%J1RG +x1V*Xi=2 DI:O89_D [M.&ccccc'حڒu*GI)-[VMKIS=/Chbs|v]d\&j!~sT Vn|?%Ji6KF>h\{Kt4@%I1\=plCKRHkmKE8e@Ggɳ?ϯ9Ϟ*#rEB#8f_'gtjz6Vuq{.VuizV- SS2PL ,l jZ:\u)Y{+a+y7j1'WqV qO~jk (j7B_O:x^w3iz9_&]VΏc^\arX1sY{LgG( .ʢO>r/ŞW"kY.H_?2ɱMM1i1IMz@l GY,~81&kd!zCrJQeI֠9GYPWK3.h(3ef:>@ݵr.fm,?`qtf\(L$ͪ$T*̊Ilac &S{ lٯbSZ:+2'b[uxb?d 9RUW5?uRj{9?'A\slU<}W\5L>b7)bAE=?:=;o $:dEK[P 6 }ܮgg{ֆ7X$Jxx6_nTzgsloV-F#NfD1O?<]d1:Y{YsC>q28'od+]Ta]md6'Lq=Sw[Q'8/$7RV̧jçi8Y͛eq=us7|e|rdrzۛZm˫?v=5K 2Li-ȥΨ^negH.S4Dz?*[jJ36[ҮPqYi5/~lʀ'YC(RV*Ps%p8X!?B=Zﵜ6xK8~JL:[7mّX+O@ɐuO/x}fspWet9%%9Rlf@JQ+#DاudL<cS]$g;ʥ TCH@U1/2',[]]*$UxPX92lr6ª(C֪)WAqb31 f'aE2(uF,)j:B`RKEt)vq2S0AV,@1e[nEE FGxSZS0Rs32~T30Qj`yU+]xd.*OrcaLp9 G\PQѠ3. ղЏB5ųnL%ez=S%C`d 8AaJڬC4,z |E ȒK oe"آyHX P AO-*67[2a@Xm?&C2Al0`&T2 d c 2(NEB?"545~ ,))5AeUGAaq vc >FlU{C%5a`\ !,J?րHh@ Lh<44&׸9fʔEr ykJ@AGܨ5w r ``ξB̕}Ρ8dR0 3,Q k175D&9zR AEE2T3Y(pseĸAaڽnИ}w8K$-`I'x__b="G#ށ])w A/Fs 7DR> ˇ$ [265e 2Z  Dzff=*k0[:{(;,,fom+ߵF wmI+lH}$ 6c,7Ч5E2$eo OEZ vWTWWu” z;:Q҂6;f#5e b7{ 3`*etG! ;7,TΜ",њ:(a .I١unv96G=xy'/JrBU=&6xH,!@# @J$' UI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $/ T\dvL$Cx K~@# I p@I $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BIKY1@ CAOѐ@ O IH1aH! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! lI 9GD ; TeN"$3"`I $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $ t;^3^%մNOӄ!8n;ap՟<J}L(I.ALрK .K:W .xB\59sU5XH+);tsU34W,#2W!\q9sUHI zW\I©YCyP?tDC7/ }'7:O8^Z|Wï|0a,g.hL4T&P`, gt);^ͱUwGY%j=m: 7짒MJv)-?v|܌~=Nu7"H8tM[?V =s!^+=.h_d=?VS1,jc$Be-e7dg*B}2L%5?/X0K}bqR2=ߦe[n/uxgPGƦs]_!3)~EW}LjžߛݰOeԽ`_as.tCS+w6$^}9mZ^mN'*A,.EE*ö4Uk7;;A~z{~ߞSf__\x.`:$3>DEU_ՔU57OնVO;|zhFiX>MCވ1HĚ8^([ҵNV">|P3ހzvo& 6N&bw+#1ˠ-vby2FlieAH%sDh!9W%n}H,337GaP%WGf߇B2$'o~Ylo U2uB\n~w{M(뤵H?u<}"-V ˩X>u庾H+5XY6trm7My6hȾn}d755\y0&q 0ێfТ}H.Y 6ZUp ]uUs;{wެ ;+jn[%/n]t}ܮ`Ky3emG_bꎂK ډmjΚ?u==۵vTclV !lZ/Gn?Ml( 2Kt?A%m0i#J+댩 U^#eҜ2 +a[~|ꅎ"YMpFg_N'Ƽ"q'eȪgK)gumA<8ӛkRpjmk=OyM[f78i2qG2ܵ'򎸐Z?΅ͬzuܤ%> E!*@mfUblhZyiy%$KG^30ahCĆJ \h30t EQt[6&!274`Ʌ4H҈=O䬬; qb@(>$b,ǔ3wR +2SNmY9 $ېd6~c [*'! 8HmU^F51Ȯ-qvaT;]C[ӓ{^CTmoS۳_0>0\?P։'YH4^*)+-i2*s{b*M+uֳȕUQ*`ڶ2竴f [ӌBѲ-h Lg|^0]u\ ׽x[제Ӕ4IFuJ('`r2I/ق[RV-68JY@7`EFb6!4Jx%f6 163a*IЦeݚ8-Ry(V58-jM9z@hF9SҖ,ck[S/L"MN;!j6v=,vIDRQtI3A<,# T'\5qv:/1x(5E-[D-m )o'eI<ސ+F Fǝp@lǘ׈f%8{i %DiqS<겈vGH*)y2; DŽNS>}ek춈 8uZgkZr_[ "m'Ij( $5>6d[|#1#x0Leab#ᷣ mIpDُ[-d%hH.Jy,g$i =3d% ē!k?>ӷ2xlje*+O TY668Ef!HKpkisN&JZtЩ,rwm$t/cuuWvsfc$|0i&׻"%!EJÖit>U]4.\j+̃ .Cpl&.ahx^EfEuM5ybjfPaspSGaٽuo}[c~ܟʦFt#$H8L;༎,98)o+$:sETGt`l%1/!%,)>EY[h2"V[O gM (3΢Q ޺mT3_VZǫw"۷V5UJ?Α{BarbS`|L.@2pC`"SG4>G *̥DK?ea8Z@@-/re2IF#COQ*zm㇑ȧϠ^v& 0@m,`_:7ˊ>J_ |}mҡW кTӚͤv]W@+`{(LDKw tIoW2[!N:d0fm)ۭŸ3i,9Fg)ߧVJe,O3b[:~UrnD&l\ T @N֜uǥNI7Ͱ47:]qtݭUݮKuf-֧-2fZq3n Dvj/Z g[gRNJA!J^dXXmΨ3٨Pu$?nCYV{ZJIٜZ ȄAqMkcCYo}!'4 Ʃ̽:;svfOI`wuKoٖ롎\^U ++MPBl ,x&>$ P3N[Qq\_LFdbT6w|5_W≯kl27iZ7`|Z›aRe=уU*4WW"^}x%m ` hxd.9(ٌ2 c)qp9oR2Iv J \邰.EBKMTƬe,k \QwZqH]-!+,Ϛl㪤Q( QL4ڨ}Ťh? JQ $*lRkVɦxU95wFig '1D.ĕ2GLҕR\ G͠sdd8/>gLj2&FQ|~{ʜ#Mdy9ֈhLt[4bi(Ѫ$xr)]Ё%U*#)W)،.j.H/pR0VK3McG>VBO< y0s(_E)8_!^"%3Ox^'=Y쐧D&2?/n,up0rV9eM -W\EK'Dw28a 缰R Z[i}kZX~.S:Tq(d1Eڶ."m.dK| /  :P2M^9G`6(ʸdx [ٗ)<9+8j% TχM[{OKQ\hV;*;U"c2 _~0wg}k6uY5w}?e8o׏ɈV .սIg~ i/#WZ4 m:1"ѹCխ͟PYnt0u3e Ο[Xcd6w $}uZU33o}>/wtٙ5~qgޯ>.JIr/rk_+ܕ0vZd?."+~^yϡ] cCccZ{͒k+L[:y.a=ƽJ%u:[tiU_וO#7>i8$gXnm qs=1cs]X].v^D3&wct1vwDZ-_?; 9l\.n!! ܱHMY~#d4&>My3a<&fv}mͅf6~sϞ;_<E6\KgS2o;h,DFɼecsi=KJsʹaX(z}D-+$L)Y,Fr*(a|TH1RD[U1Em 2C\TFo|]BvuK'u)/Bme8Ƙ C_VZf|?Inen'X,&d \WIK8y'gӬ[ή Q{"$FB:6CӉW.t_yXy؜ f:\0;>4FooQ5s9x>dNh\a`*r6lhu6&OyJrbmй=RMsVoi=[ H JZ)TYHۈ1N -%FWRV_.D/C~yyK}]< zWʯ&Tzߦ\c}Gcի\_+|NEj#q+,(~(جn&zMCWNе~ShOGtC}`F9KșJE~.5.,ɗk~O-^Obi6uþσ2*,)]$1_}!1qD0fWM-"Q^ HiZoUMM_} ;Me7JV^*1 y0BxnD,Y=:6}3-BSa q8- s [WWfB _) sIk_>KJ  p\mA  WEJ;zpŵnJ \q<*ڃ+07WBk)  \iEJ\AB,Blڂ".?vUH:zp%4NsynR"zWn4.Ӓq7 r*QiiRSFϵS&!wS ^ /xG=⦅δq#I+~H~0㬃urF.kxHC߷z% Lx.d2|ͫ}hdWpuin?p19>X0dhGzp{37I \+11HVTF-7LIMJ#gY5H1C5wvgZMSEOSMt-:s{i:,s|!2Yf"Ld,3e&Db#Fϗ9zȌ0^%ɃeZҜ=R @8P eɆK2 1Rc?}{ .)3*k1=G+ *З~ ,Rg2D'HkD rQ"!&ed Q tܛ8Yo=p7ώ?&d;jwKw5<0 }(XxS5!bt(($f\A#3^hZssႄlx1.>Q/S]~x-j40ދqVc'$ :)5ȃ2pp5qN%gh[$Ɨ:\(hfp H):&(fEH$4"PH&ֻ{!}P)iR:>n2U'N 2E{m ɵIV-†QEB:[gM3˱ˣsMM(qQX NIꕳʨ)I[iL0+1?YUmA0T $ G."F5() !eg>x"(|eP"( &v~=Ҋ "H *Cb8R Ri&*&,%|5ѕ %e(ƅL9hC %GhrZ}eB`@oA}>u4qy#;J:j%xbg*-Q=2B2 T[CXeGs8֥qa#II׉(ޢ# ֒"XgFANycgˍmLlqu}WJv{gcʝ1L7:0݁r 0IC1tЕG ^ 6sc̱b>bMKE$"_hY"MyhyhP0v3SVlp6HF'g)Oc.*SQ Xe3hj4Lz*͐ NRA&I̓F'\% \KpOo |L6N_8qPG}"/Rצ|w2 O|2|w<6%Rg߷?~ i5ޱ-xd sЋJ2(HZKIϪ8 T l` 4G0X T4)zuM;ý&6Ioudq %MjHbꏂ|ip qd0Pɂy c(^Y,é/d6=+mdtySm  <"EYR8N6(L82(:oYk5f,`ZFL&Z)dDћ8նym:f<]V[ruJ38\NFSo#صT&fճJI6`3f]Su3 o^d:l2{٥k~rh5-EVC0$mnTKvdy+'ھ]vUުpّۚJdE}dqfWkWTM8O@7?wzW/fmMomv:.gd |™Z7Ф2SR\-&\^`A ƜFo؛8-cX7[m!ͶpEBN 5ww~]0hXΟvpC܂ӔV$ & ւAi-7G .28ZlJA7$gxd65v)2h/#vHHLBDN-vo⬷v2K.XFǾXmvFF<1F*f8EeAS1"%J,H#)xm^/\H☁ 3 l rcH)y{Ҙao⬷S&ø/E=[D-b+%`jNj h{fb#4F`߼Zad(QHLLh`g?YoIǿ tya ;}Y4leRMG`F($RYdeF7,?+#GAaw Ď1f{9].s1uf%rQE;^ 5Xt%z+,Hjʚ(g{Tx .#xx x-mgu CpH5B]ܝv,{klū@]HqvUxADॐ5-{sq DꥶT֎J̙ʃ^T6;EfKKpUi$/x0qx&ՀG m|Om:G-r}hSi +BZ?BĎ+TiMq%k-/Q_Çżq#V/\/ DX#<ȍ\?'_( BtBs&PyG!vW} Fa\^-^Ux>]-;To]< ˿w8ͅ2 WkYXaIa k/i?˿Ѻg:+x#ЌL./y}#<Ç|GPy/f˜rcyV3h5<P$Wۏ? }iHbg6?7vlX)vclb7·O+,OOg^ކǿϮq|9WIs#jtfyٷ~fWlQU%|~퇮lI7FN)\{ )SGY/X{_WF[E ]7KUd^T'ffn^tY+?:,˷=?/ejo|XU0t'w_|j:@Gx]cvDíb.K%~q\x΋g~7/VB_Waʇ 6d\[ޣOc޼Z7F4Xe= -"u]:S@Np w̲aH!i_ͮHk3e}zwqK]MiSp[ϟLj̤Ԋ_n ~n3+vw/v#PzH>rsWUm7uoL|j>c@sS?%M# 6|0")UvM3AP?>?LL;ϻmNpj| l_rj&=3֩4"n«w$&'P5#޷k[kCYF]֎u4w!741ŴI&6,O(e:jʎO?1ik8 QpwU.pW`ׅf@عpŭn_˃}=̇;1M⭅cuu'D|$Z(D|('^K;E@Cpٍ#T~\IiB <g d֠Z*)5tZ7Qɕ*\Zecjq%5Q4% an(\\S\Ŏ+T)w5E\)xf 6xW(ŏ=urGOY?:2J̸:멁&+n//ؽ>lb,Xd1hf23Wy4/FvCtjvUyb;yUӿKԻgK-U{aj5K^(d i[m7W-,Fr/0]kx 7e5 MCZN域WY]R67gw~lN]{_˚ @٤bp6gY (D:fc72{8(\rكMK9{d 7g8j)k 1Bst$G J_HT\)I^ EY{g|]faM:w>/MfjMcF{-4e'VGWt2VI}=_]K*vY. Ӌe=wƐ zdP? ^;^'kryl _[p<,bw?W]m(bq7Uh¨P)qpt;o10һ{#rf@l3J7lR`v<\"UIC] m\$sR*ird9!ȳI 4XC -ra9_ZĀd9p#&P\BsR6s Le [M0;]1Zwo~AW X wJ~Yv<3GlNKTʈ292)29a2hBb<,K6%8ɑYuvּ0 X!IjZbmYK[AH a|3d Aа(tz&\8O.85T2k].G1|FhJkbv6Zޞ N W~Wu"|,SQ{ (1兗%9L<@ ʕbeoUh)eV2, 1%mD HreE&Í k29ڞ8=c=RVӌC} e .%*6*lH%WYnRo$'B7w@~o2=v4AkG4IRLP UFΞ`]V2j U=6 I0 (nq4 JVd3q;|ʣ&G,ZWcWgdž QcմXv`wihoI *@<]2K VkKy]!\C&d(ђIȢ(\4Gх)"j!T[R?񸗰8cP+{Dy#\sB;z#!AhH=r^ "4&0 zr0+3V7F$X.-hg At n%:*ee #~@@~q̍:iɡ~V_ܥACmI2B#:DBEQ2S#L}F8T2:ar/?a\2s]/1rlX]̽,f%0;LBDjŞșŊYE . sW@y7Q8vJ*>ƫߠB$fnCq "4lȡAvUwQ[1seU!Q%hE% 9cYUH!,TNvQeOJe+p \+1Rx2t&~H>q\]*n4yn^ZTtγݕ[—O &xd%1f,Ǟq͵Wo晴)-`àQpR`V19zBvߦŵ%an8bom;ۀo;~LFYc}g31G%r6GkMy Zɇj09ͦx3om\Կ"r{=f/_c50Pd2FI f$ Y&&sJ` "]b&2fS7z\|2Ç@;}tcQ 9ץ)>yq %58y f ȄJ^ :r"K&Ȕ 0i%Icpo$/\\3cݕ]0ODaPΊ{) w"/`fqi훗ph_t\,? ͨX%b#*pBALF7 6t9پTG GtIә`s*YA<}-Ft*e >q릫{4~`u/q-awFlm3Xm|d9@15x-BR2$, 0_b, p[D]II2Gud__nYw X,;֒[ h+aCCQcq;m}l}Sx׳7WE*Ԭ:T4CITJD);"u "2r, Npz5:Tht⠭o \wȲ2倿}u3n)6ݑg"]^˵k4ͰOԸ}P5i 1 ױQN+%"ZٮnI[pI0A bH,`Ii|V\r+z@12<PۺBJy % MV(`ιX袥D2^M&~H塿EB}v^5iJS+͸YRYl}A灘{IkT< &<6ml|\l/jKҝ>GO9yㄎ*0E۠PRW+n:d&PnQA~vLHGE1:/:xϸ "s"0ɹr`tk-OYE(Qh*\28ݝc9%#Ad޳%՚kP(A+ud/]n:QHpL@gζG[Oz=!$/t1TwP̣HDwy/-&& <&i#CS2'+2.}V!YQb9bJǃ6+V Ӌԁ<_QY@9ِ M2̉D?2< lb<62ԹM=Njzβ7Aɲܦe^ad+$A ځ R1a;fmhf?j I6~^ċ ݯn4a|/܍'<$3H:}qI" %<MI`RS$\΍?$w| rWO/3)闦8/W'cNܳj&}X^m)NOy;9d:/^tܻH72~z?˜~{o}${a?V+b{ZkuuoEWpvx)%;?ˋƔ;%wg9wj\5zwU?]8Y9MP=^w,"=NFt<,l}߶ ?K)^%@n;r;V>ǭEV.PDzV8+xMe%K&nn~ʙn!?wfLv#!o969TQABC -\/lq\m8}G͙[[%g}i{9~=qy3!w"9zzwTl0S2 {.|d.AR{ ʨLzk,cd&nv tpǟ7VɺSwemQcʣ9tJ+23B+ot%B-f#5DS3adLA.G!'H%eMԙaY(B*U qg12TY?mȮ W}RWIgB-R޹>?gznPGw6^(0S6ʘ$nSus|2k UUL)M߈28/ER.;'g'~ҿnSE!M-X+G3IS{aɤ% *ܢK.Ru-8z38ޠx 9dBhRKloq`pXpu31Køՙ% ։NMN6eאf݉vsA|'jBhC޳#WwC6 &3X`qA-Ds~nZHm[t["dYJ^Ֆ>%J"RʤH\8(CʘNUDZ)! \&'ud,3dl4rٔq+?k3pS E'(_)6&j̀h:P鰝y972D G]Jy,W1h ro$E$d`+4K6DZA]c%q4]('F|І)r j|" p""HL%J!9sҙ%2E"r͕R85TG a{D'BA"#R;iIg%]YQVw2EiNNk,3 Kq⽐Vj@b 1N ؓ@$🟀KF݂O?2'op-0#G|O;㹸S(,.YT@zU3U/FI+=QhBHNmH&U!y>"*D+9] F;3NYv~ĕ |Xn#q1 Zdadp(ĔРb0c쩭;v3E=?PX⃎.uƊ# gIB[C3"hB$Yf=(P@DOݑJ+dz{.KcF'< VoGOeqH: 6y`E%ԛ0.rּ#\8+h,:C J|VPP,q$o'=qWvڴp:-b昁N B9yfPaj8U"^YhWx$FB2l!qKVVؽ˅)W4Niė>3Qa3Gr7*8w`Hf8OFNWTLG7q!"Bb%S8LVtSSې{Y.ڈXTHN:kSHDɕ0^V()5έȜ ]FW{xisʰJ6ڈq_ssiZ/reˣNl36ˤ(Әics"_=j3?/wd[#BzrJUGR ;WJ|r8_n{ma!с2QȤd!d PK(O@H蕹>+G\r 9@Fμ22%3]T4,#6@p SIG{4.ƃF&p` r'ZZ?FPsF~Ӧfɗ[iژϝkUj|_vK>3dV`*evSW)ќw -%ێPO+!@5¥D ``Yڧ{T//.v]Eu͢u=@oj=HZ^d \&TCm:2Q5hk(5&Ӥ8#LJT(6F.gN2C T $w38-s.IölhoP~UGx iY18+L HDĐ#CQ%7p?(Eh3dTH”HYLq #-}`\hna>>샙_\')!/Vqv7cLfE_Fc ; K&_<^|L"L|VjI"7'$(#ԖdJwJN;Āh·jCc2,e8SUrn5eA<mz|fqŦr%0 ~hxyU-3p!Y%/x8~ iIm)O5u0y<&txR(\YOƣz1ž8*%G'4j\k3'<@ds7s_i=Ocgri2npsa^Ϯ5/Nbvq+\캴!/]3TDLp\= //~z2~?_o_, "S ¯ 8/?XWCxL6C.g\R۝qhWP:q@g׷m6Ҷ6u^şPr&`Zd99{U b<q>QK58нaBUBd0Fpar4mdE%[c,/}hnQNܹ$T fsYef 'L|Y$L:/ЫljaBK0c\⡟?Y](ץ&rZ ιb6bLPsEP%xᵦVDC}@qSFs2i Gc XHE)1AX'-ɧY]ě'y dQs>ݼ˸_*#TdR!]^E9E8s+HsFƆTl)ף,ۈ7yZVeWрvuGfߘҵ)l{M4-yXөu}3O~2i0Q@T6))vIW<}3AO@K1xO1E'GtdKk6Q8B*Fjcpsr/h󜠣<-6+Bν~ |:^s^l]`ruQ=.x~Uõf!,e[/z^7>yh y~|M&{=[Ϲ OƩ]Dۊ8x]])L#1=Qv+Ԭ_7mͷ[lSly,9|&,9%giF oQu 7܎|^+#/ob%Z,oLƱ0S%TbS4T4K}gNO嗓1f5};nU?3c-;Kj*퉧c %6>*gkvP].RfK7wtP%,nfyŢo^*}S$dl{7}-/#"ү@&Ҋ._UI| =" + @`o^Ӕ4B; ))LPxNxd.yh;̡6 ',^tǔ\wM唭]O i#gr 넎KlyٿF?cV] ~5&PV &>ď~F5= `74D?2oi NVzlDbɕ$a~"jCfR-pVocUfl}ܶd":*}lo;[C*Xɦ;cin5@ĤH@ 0A)Dl7m"c8gqAh#h Ts%]Y"\_ߌQj¥Q&U(h<PQq#"p I M"j1Dwb \:B1Q*u$46yRpq{IYK JNF"37~W14LBfQBr s6DRL8eM4 vj)8p6S>Z;q!%7IY$Y Op)1eКQ&,EFNd#Ïdv!.#Vz^0AJPȬC,w%so5gAKi]䞙[U=N`GlEsmOI=.]FT,CD ɞ5)0sS@w"V;w_uA Y)dW(! Q S ՅBݎ@fɣ^f| F Uۂ ָh#5>  ^Bd` 0` iaP/zV&Eυ4SQ3AnPH(jO^S`8%-Cɵ+F(_uF>\( m TS.XoH!b!C1 `(I 袀%a $Ti`/OH]gF@ޙ"B0@_/f-n^,N.쇃kZ H)OUH 8WH-WL`kaoU_mͯ6kiۡ+Gu:wK ^EsryJl@V'BVktEe'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vu[srEo<'µ8mPO #<;B'@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N dݰ~FN Cµ8( @_CDv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';^'gBi8~>N C;e0@J N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@z@Wko]Z;?RSj]\\^PVu;eqzxr9\yG|pc\:=? ;tXÜ]G ʙÝLrju;i簽qݥL۷|,"jjo[  ֠ӂ~z}G)oZuZ.NK,ݞ5f:9>:xpdPh_41o$NBYt[φ^;>v1!eSt|hEsWGG*x>ȱz+KSWMMF:EoTBP>j6ÃA~잏вISD "!oQǨ H/ WPv}{6XqȪT$ #"r(j#uIUU4xWZK1GF)yK"ŻK97;Z,#_7v?EA)漠ijӿz9uzD_6ɄMjvm6,ShLf堕OWr]&r˃ze^^HTFӟWhp/j˓mOsCDE'ݯ]SIi q}G[pqk.χgo#ˤg#@Jg8N'}^igGټ&hQ:𷷏١H>q}`n_*嬏vE\5?2*u]?uZx%>]wi-;)Go)FqCԈi<ged5]>0ِ`>w `py|_7{*yzS1N4evЫ͛Š}_9}&r^W`߭޾5np`7`}o_noǖmvХ}b~|z61/Ƶ?'?~!ͅa]7{{jZ5|Q@Yܔʡ-*sݣLRȨ߬f9?An/=z Ռ/^r(',3v|OOQ7? ɨZr2mo{KO'os7fŹ|\iv9=^kz1ZTqĮ]:C/^ͼ\=9XԛfS*-r6bW/uVD'| %x:13" mWVڟu${3;]cv;ZK_S&pֵXN^2]MMi?j$CԓH6Y%V=spxzp71w%S# QT, Λxۻ-,+OE ۜ lсdrqTM* P8bsqX8ƙ0OBQ*+a͎Q.DRo]l 9['jKw:yjEC;B1ihiTE1ܺraHe7*'j{}<zvO,뻎w jÀ1b,T$lM Dw$l19㒩9z 1l=uy(b(un :g&q=M*VߡD]Lt|Ň=,;/TP{Xv<6`ox;i/ $a'$rpõ%-5T[Me:JV(d0Cih'<6Qj*kqͧ4s;΢~FTqX~<`D7>@iluHhb#P`CL#jbM:R H-Ȁ#vu¹߿i \\Nu˒qqO:f=2G6(!!waJhFm,댋 ˎax |Zm5[>GohZY%tQoF>龏笸e- W(浏wXǫGzՏp9Q\%?fdme2A)OMѭYJ/u :'Q?I,u9hWYZXnl uFHѳ&k!c: xȧN¹;Owb!x/f9f>,|Ƭi\%L1 B+)>Aٻg& iIOY49)ǿZ'ٻ6$Wd#_t]^FTK\Q$CYVr$JP25(ybĒg=Şꧺ24Hm!$*2oȢrP6Ȓ)44} 6%(aQ-W!F!oLa>6j͙B\>x=b5ukՙ8#,+VO(m8eKN㳧n;C>ւUsI%2WFkSO ;TZ꼎F9ۉ.+ ߦ>ߒIdn4+mjʅ&!z+Ov.+؈>&x<}U;~lu5Fuֆ뙂. >ZkU-;x!ʤ-RRW6^s:[؝k#Y:G1b)VL 0e@;LٟZsHm$$5*XC|!0h4BItu&-ط K9%Y;!OҀ51kB Kt3{Z _ Řc\V[+lϊL0u]uRPػ8?}盪*x\MQMY񮠄K}ucq\TCIA#wF}W,fV_{o튽{Yb塀!d<˥ o^º(:َƱ>~> 8vM׶}KwVOÛ>$}i ^PޱzqH =L=,d?Xgx[Es8|tQ,j9:* 8)\׽,ˍob^,}6Fi(@\TGQy<·Em >unjLU_Ws8V;Cz =*j@p]PiGPkza#;5UY^O : ; eB'dX&tTL3?ܯfdЖ{O&N0( % ˤ$HZpl)fqQcs="}c׿Q=P\.K,rsrYɪ9a?< Π`l%z8ɿA4X~ /1U֚wŕ(c'Nswv}]~ȝNB}C4mU"t V{j^ihnϦI72Yt@B h["%(#ID Y x(4LjMeD.45AxgKPЙi`-jKSDt5jgK % WFm5͊^F8$(7Ll(aF'e]j筼׀vKʍVDx٘G.Ny#C J XJ*QqQ"]::nYTRvƦ"[3kC[Wb7Fg3[5C|x /Ȣe T/'!myOq;d\*R)ޖ\y ?PíQlqW6!>nKpƨ!"RyY ("MDH<a U[+}=@ҍ}}]5/_v>P y ˓ igoOm~n1:/F-Y/Dw5.r!k\ơ#|yE|qb-W(J0\U!|[ 2Qf=-O.A5^WN]X$Ve^(\eD&s=;BώGjH[PYd;N\iڙn&,cS>RlJ3.> 'dXI]WJ]+`EMSK AXSĪP_^{",CF'8C}\ci@LW 2}*WehR2qAKJh( P`AJ(g(h寊%SA SGd JɌ A`Jb>m9.(&]by`&?ٺb >jc8)*ɬi1Z=h>4eNCYH`G3 ^VA0JɴhB JIlaLO6u6[yzt!%NPcZhJI{xR5$EC QGK@WB$Nc3>ݵlf߁-w ~l4'v<༃DC:PIi*иBIZ3':X_k#3Nb>Z]Oy^)mLFy\!y9a (42g5ёt 0z|+%wAw3{o$yM -SLE&0NǍrxVr-Z(ά~O9zZCG!*Pud7 ؐ>\ȎRN<.FYܗ+w $2 K6$o:v.("g9C?&RP)Wz7 {Wg)rwCawbj|Ldzqkb0|Z|4ÅTCurslOoᇣj6h0|՝(Av׽ɇyKi~7\@Zk?? &2 %͇Z^}w݁e$')k\:Npִ(rGLj"<34BL?o^̩fLwvEǭUp{Ql2 rk_+Ϸ6\ɺR*".>(p!WrM=V=eDP'PbՔ"MɚrgnNޅrRHU,T-rSI]Ȼ)%<$mkU'ǓPMiL+W-w4D}?[m`uɛ lYELFEZެhwl6f}F&ev'_4ܲHa4(`Ԑ Т h;oQwmÌwc"CwAMx`WN/K+Y.<%MtOΤJ\"ACRh(Dyc "|EudT.IK3ݬW;J^_*QGi|gD) $TPSb ),c͛:Z qz5:.wXgQ1_Lq;ą;^g3Ydզo]ƣrvͨCIZG ӥ,+SZ 4QHxܧPu3(PLR$%pyi ѿK7M% Z#[8R 2\'.ObX?M,uE}EgDCSeV#L9i*S ZK)vkR**΀g ,=Z3&1rhT_\ו~ ^^OYjbgڰ,)Ԍ;*)^ L6Dny_,{||׊eS;Ų7(|bMAaԖJ}J1JdL#Bj2F[mCī#Z A!D*yesXtl b6)ͩC3qهRE(IE(_ %oK՜QjjMݛRϗ;ʓaD0bΌZZy Q$ TJob\,ۙSgفU!h1RѺ\.X&'M,D Ѥms[;s$9g ҄%,3%\4d4N8qwlg@;W`fMvE'>wi@&ycxt4&M^Y^rHYМP B.&Х |%~~fO 5r)~p&|H1x2#'ZA0qR'=aq$7 -B2'@=lNכ #5%t@DkmHe8@ #u}0>ǻ kxH,+W=á()r$Q$E4Y]*G+cF[Fozw>{ ^DLdJ(GL" Q 1=Xg)7V1ƍNv=b|Kz7mȶZ|0|XqDQL"ipQhk4qFmSPh1ˬ*rG{c~8ڗl;h˻D'$ VoG)"{av\2$R$rI?XJ$+K ]HZ& mHm%cR"@3cc I2&q'ɹxp`>QP%a훩ug>ȅ)Wlҝ^{dÎZ ;^, JY.;iTA@3`ONqO?-k0@\$W&1PB]ؐ|@O9WD7{C]JHom=mj4SEp2<.haBb.3&H {CNBiR/xM\gyfaӋ$^a⧋~xE[`CIX.K!Vâm w^`)}dw?Q誱͠Ӓ[ڴ;?q3/Ռ|RQmeD(iX}|}T*ZP RJ/IL٫ JK;u0 bB:ϒ1E)G40!)MvQT6F)D; &x<M$&g >9CS[֝ 0j-` WU}5>_Nv@lfVjxAkПkD ydUgbvMS)yyUʃO|zwvNɔ{WlV㊞!v|: .v.]֥[DE {l^;5j.Ff湒wCX]_v=7 mE0CTI'4U09.JF`Vl )+M;i KYtݍͥ//-SE$BgSJ󙖣 "(v1xZ~ qrGW?C7S~5}.1HFTQCb~K5|ϋe9I]~rF\(o{MۮMsI­Lub˺T$ɹ>?@v V5"ǸL ?IW}1A >C(iJǩeS7cosVl .k^B/LăP+p|VU,㮹W1T[=ޠ~]7|?O~g|7( p]Σρ0v]: ء۟7oX[Mc{MLM6mz:KyC^BB4)BSici#&e' /Nr&`Zd pf|׍A Jy"62| jp;; sC~c+\8,$#a|Wd(PA\T_U۝lyWg7}γiYtYVu;O\'"'SĜ#IY_,}*7l "tc Vr' RZ QPGM|B@(nzZ-"@(# 3^R*_Gg4K ϊ̧xmeN-PPl,HəF-hM,(/$FRiEЁHO[ΆA <{7TY w=S@ Sڌ!z\y;”GQ"%N ܫhhG ?u n+E3lhtÜje VO D1EJ $Yq:\ 1PhYk;-$T9\FA%њRWyEECTN1 s$cHlW?S!["Õ9;9)@oP(J_\d7?r0oM&E῾+&x]aB͞/°c?Au;Xi'Y}\ =5'^T7z飺UO+=g8a޶B vZ( Y:jl_&p _ͻ%N ~h /k턑I?-&5\>ڌ. ғUpPq_T{^ͨsaz٬E.f^ϸ) W7kv)’џe`F@3T`F"!b)] C@A B@vt;%cmK _C:ک0H¹`Fq+Oןc1Ybrrq^QM&b5iRiIq)c|7WwAޗ_{U6'TCl Ucy#fE-ꈤj U.pեgH"q?Kcr) J:p@oK}rٓљJrN0HSԂUTI% &.7:N$Gp&86FEiL 뙱%!ЦD-.q)g_SZIJh^ֺ^D]BjTuGI;*Xr%Y}$H5mS1R1é5XO/?p-(1 "RK-+@Oш)4([i<5, ^(},EreQ"Mx++0` N R{"xb҇UL$@ȗ_lfRfVP8P3,ְZz yx(V/+. ycd~Tь9!6]u =SwӛSSH+F1 Y$ZN v;޾h::g9YBAKZZ %mg<}W+llvY4N0Hoן \T,{X?yy:tG=ɾgsJK˵n\M=x?] :F;ǸRP dK F#&:A93'=?jXVfoX0΄y$݂ S_C NSNZ Q (#WmR#苫xrIg*W_%S$W(^%q(I4qf@} pyXfUJ~9҂MсWITy'v\t6abՅXhRDP̹9g;C0[N!RcS@q,a_x_qk|-v_Y{@=@)Ę[nu"O!zYxF3gvbU*g3Jl,(Ȇ *I0vKA8)ʕz@+L'-c 혓‡)-9^9ERjl%~N:ki>[rʦx74@@J%m]R+od枑nÔ3IGd妳)E85έ?U!.lu D,cf)ӚxKhlNj!BN܀Bi&2+L2B 0MJ6r$sVg#|3`zN-T}(BWΚ썑MaZz`!rEВq%0@&!dχ.wOo%zNK?8.CJ )F1xn1#SZStI(])%EҀXL%dsHg\8b:&H|.*%ښKI-U5]iL&Q&Ȥڙ֐x p睰)"kW!.SU΃EOZ["2oe wH̘ Z ȼh\z / "3أ=X*cU/;Z[!rh;Kt^Lܠ jRMyPV$J(" 6y`c[%bkRK:`0PFWN*aJc΋e*0Kz,0`3FИiWx &Vw$ o`خ>r!VKC!mx59Vq[ W3p~T޼ vӨ>1ihpAu٥jH[9$~R'-hCtm%WuFӫ? G_ >B(7W R!B46IZiCQIbk ܯo%Sl[.5Glܢͧm?eSH e)U%9ҟL/g啂>M?M@eR(МLrbuǡ~p"6I: ڪhj0'-I沍B3zlv=&hgkm=(z:$dUA&ÏC(bMgA JVNal5sʥ>^ E$bqzX]epgmbX4:d؛w%vsw8[5[&[r=rNޅO(`޼M]l$)\]"m ~:w{fH3!c`^^z 8$}?gma:y`cnnhn}l8Wf.-Sޘ: ]i>9P"BZ:Ok!&.0f 2:(TP+Y2͞jl9\Ѧ.WV`'AZhmmg}5 xxzW.RAS\8% 67tV׮GFjzͰq#teQ 2)V& C>wIPtH@9%Q!)- 8'pk{Ƹcd+ eX໔‡H)|@E],[Ij[M3tyg!Vbz:O׾.4G^~&O%V[ B3%VVU:i\ro嘽Ϊd9W$8,g>qM&'ǵɖrT$_ڮ9sF=h4MZPؑ 6EmPn+:,KTF-^%꟥DV v !P+(*"G˥pVJgeIZYP=ii6[jYkðǢ0FD} TfJNkg> )0Ra)l ۅ 55H4&,q!Xi@fTL{ $X yEΎRRUz^.t^˕' tT[:`̚yPg)jұH(E$,3L&nT*n;f:_[tG_ŗ XVd(M96BCY@zL<9bT24c mHb!0.$b)rI|Hx*T:1[O'Эښva$h Zrі=z&x.u:DF rDxee]F'$('I戊ET98I!D0Rzad M`Ƨ1CV"w,a[u7K%1qĬ4/8s1-גNҴ7RgI7y)92Qa؅{Is7] ;%[3@MI=§*-gB[{!c8eKzԘk\#73?־Iq:9@{7L_jիO7v wϫ%׫ K0rpsvjE].ihRZ!rh?Wί ǽ+c͚Y1c}t~rvhVĜ<蟜Vܮ9Hh ޞrri0Г⢱4nDs7bJpT(2e' gWdSc(V"gRi#clYtd}y󫛵gvim14+|C3͌[`[no@?)4cޡYbn?7gZy(Bf40jC+I%\; B:cZ!lֻ+!9e Ybx;qONk~ XMpF'7QDcNʐVd3$uY~z#Wg3?>IM0Cosu`Ly-牮0fp/S11jqG+oX##f\8]}v(JQjkVH 5} st {D { ;2%0t J,E'|{F@(X㲏 BɦƧ)a&%e%7i-E " `)BJI%\$[<$LYf'H>Y*H$2%ͯ"8u)4Xk#2z ]lНNU; ɮ'#91xP\PóHW, #ⓨYT,W$D $74k勒R0\1dgTFH_i"~3CȔ1YϢWX٨R)\B'U3fX,cW_( B^}A}᝽.99ex~|3UՍ?uw>|;('5рp4IF:%{V uKeAʲ!PJ!=3βDjr٨r;blb‚*H$hScfD1kcڢz`xS>&Hh>]8#"0]lpi-ZelJ2"C*E}MfBYcL2+}ZlxxbG=z8ShRR+l@D g^1j8u6:cM>Ie䬡(-ڤ4QE43>`(PTs0cBs)LJ fx}x:Yɮ~Qx=ѢPLu bd3624a*Uxx,u?ԻoO" Y%< ~|ݏSbcoIn4C,P<{yuOԟ7H..m9%'X^魘T.~|K4s[E./HB0vş2x,{=uGE[5 }/۽::;hAWBDZιo.k]BL-ޫg!gK#r2[cZ""z-N2sB5}zY{63Ǣ;SF[-Z`@c=gC$NW۾O_ qo QsZE 5q3?n:DerT/+&j,AV赴)A@INzАc1#a~mԚ3̡Y\"jO8KNPXbټ܉>qCsݚy~pJOL8f.Va!j#ZLԩ /8 ;4Z꼎3DE:x يg"A0{.[b|2I;(. |[ SC)9kݤ=/Wn#BxyuzPO]vc%0.&2 H%傩DU 45@Cib܏2mv>[c2\ו;!Ι-=)ۼv,!,aPѲ,A,Ld#@H.yR𠢷XŽW{'nGnA(!@[)#Qʥ) v)5yF zjaWMGHN(| Z4m8g]UbՎ#WG(TD)ON*hmJ&˥Ԡ!1io# #W-yËSOY{<}H-H1&-| DJO E18t`HVHuY}mi0(г@ƍN>MtH΃Y:ƺoJ+7pbSSag,x>=:-u i+bsA!a=m=/9fRcHFXHݎuYPVԚx') d,o'z끂7ZiSyO9<oK;}ߒ1xKֵ-ɳᔳnЦ vjn^R3ּ=F9rt' L 'fG_IW<.c8XZ\O5&ȀKI!boߎU~ŸWSGof/n35fE3 ~o3BM2T淭,1k0EG,jйeI(z2$'oeBGnN$Y'г79 Id]ÿL Iˋt+wE yx3;M)")<@3!.MnKvtӑs_j:s7oD8 FAnCK.TI{WC!6W?h˟~yՌ"]M:Wq%?*7cY>&w;~h~뻫`4CNdw*~WpEԲ=BGnaZW739d3+z-%DQ(RA*gNoN$)Hxm xQH\&2F"{Mpolg<ePЙi`-jKSDtTZȵXu\p>%{eZ 8{C"LkɘNKL:hm" Oh@rH Ko,v[-H'<$|x0s2.HVn@U˝UgmcƘ\F#(`)yb㜣,RE« ttܲu!1E(3cmH[qՁ{.Êb|uH=l\(.8ҷ@ehJAk`-$ɽAl-W4[.z D"qh:|`yA(IQaS Ȳ4.%eKuؒ>C*]o#7Wm mM0M. ${-pYcG֍,y-93 [˶$2m˞N0ME6z9'V\;Ц9#K B Z[My1b. bh[O6|"#khD6y6:KLyl@D. 6J7w-l//cA}p-g:EZt'tT1.J^j}QePx̙왇cCǴ&vwKp:q'kmdҮdp29d5`0:Z 9H\9 M-F"|=d':x!ghg/2lx⤒j͵*.e6,}$#~Ix3 )P{$;#mLH>l<:.ݭbs7>k(l疮&01󘤍 M䮈0BV>YQСFt5c* $uc$K<* cLRJ*4>s2'J>oɜ!ƓAM,uV?_,u"h8|b칞 +Y^i5#L,-H~:z.H-Uq>q}5Co=bYB5<- ;ښ]?N< DVK5xaY]Ѝf<px %4ΣH .ʡ" 1eMҊKɥ.Y=+Y9+ֿ>n~i`Ҍ^ďNc$1޴vIlSJv˷qeߟ 8і(5b/hOwdޢw)4zBEE23 Ozyoch6vUэ'ܮZv٠8m$|#?\1v3:M`\ @ק _f4,\XvccUۋaAe),?ܘlfGSC3ʹ'ظ7M2M$u~Wx6~܏O~ξ#;=ܖv8D7[<?>M$4ח,߰0c'LL']@LGmf&;5oj|[i7Y&)MkhM j[:6i;XΆtqEѹ:l;% @!ɣ3j <]hF_"h1@}\39pG#ԖE|gca+yyDާU ~@12TY{$L4҅l7S"+ rsE;g:BMҚ* QmoTjB/ ".{Y2+ ^ړ' x׮'a39 nX!:k ۀQ,%m1ɐ.OgO3dD%4Fh1:*Eځ1ҤdF9"ZY* )ݯ%o|?IͶKjwSS2kumaECG,vm%DŽJFfĤ) ,3A2iiw $%h H.u989 pnGC;9 EQrkФKyb*/vEZ+q hץj]{%ڱ.,FsP5K:}S9SbNlˮ! -$ȮJ=?Ew(zCG=ѐ֖>g9cLvs~NI),@5Wq*1&Q&PJ.c@M"fVAyc%W/ֆȹ"Oet3MԺ f.yr(ޭNbJFVt2*`U'3 3e&=\kVn^ "w[s.S5L6&:臀NK7Fb2̳Z WϺ|Y ) 0NS6cJph-0ibY;\vy9[YW6{mب%sOjJ%xp98') h5bJy)ͤر@Mm -,hԁ֏ˈ(/rT)n,(3'AjP+NIX>Σ$ HIK8H y %y&M(#rjlBog8' *cF퉝FZ4{e|;昦dB8OL E't=$8sA2^b R`:ng!.α)Q {]k|2m1+43x@<0ImK,NdI?F QD)mr; 0}j;dWC|FЋq9X΂ovDM"$nD?;$t&Ol"Ѧ,g5n5 n(j_ެ57B| 3Ȃ.ta1@rB®w*auCZEXRv&k tH獔 /u)Oz7vJ'> Dݎ9Ts;^ p-Y}v牠 q0 p:Rs68R9* q`+!Q9yCJnTdɤ+D:4t4>rCUmVΥQn]*UZuJ]e5+^^[c3I&`X.h&ԮLHؚzߦEne}Ecݣ>? GklNUF9nZoJY˝جθLz2Ŝ6>MУ5d͕ur,Lc-dn[plZX A'zXS?YG+tIǓχ0it2Qrs94I1:0erkV.wcb Pz$,Z0 7Ձ ̀V$gHꐹUAIAt#;yR'!)t 9I KFVjlSV/ rz67z- ,ES8ޔ/Mtj )]1^}"j0%1?+)9@kMY4\JpE}+H%IP) lJI(njA1";UJ&Drȅ`I0cB BpOJ]RY,p,ھ2BS bËGLJk˕:wr_0px4g0Y-gQh^+<3x.7`m(:]hf殝7̗4) kPX)Ge*l'e*;PD[XXen ʏ*H"XxuКT7I3O)Yf_*qo^7a~dt|Aeiҩm- Uskڸ3u;nyµIsyg Q8>\"utɆ rHVBFIZ595P^w~/#|C +L9"K8L2 qϳ hbH$ST1; T ]=o7.|o]|v X@*c42e!FJ4SnDQrB)FZ|gh8eu(k4 k]öCZOL/={z1ۇӏF| pi9qF%\TaObSyXP{b@\2{&F3+RFA:@LU_ZD]koǕ+~zz?8xCi&g)G ܞ'%6IQmK. [tw[SUٮ9Z͆L+)(iRuQk]^e9E M;9ܚJdqR̗~Cg=\??o '>ګ5 ]i7 >P 7f)}!]}@3=u6^(p #ƼY]sۋG}0Ft1yi+: g#2HD.J){tbd9Oww]t{~.KQQm lǙ _wDٗԘ/׬2/fGmeyYk2vC)n>=Z2E<=1hSĴ7f,1zHM0soپϟ}a֮vɛ֮0.˜а2˭ l77ٱNֳbG_m0h q/4>\mnn9]6~֍kTkqHwClҶ]Nf߬~:]օl<޾`Q۫9xrRroz}ʚ~K6Of5ܚzq|}J ?/?^hsPhc Rg3g1"C)&XXXǽm@? :KϾ6h@> ~*w+i#ttXs0DUoQFğN+3%]ꖟP!U w7f ysLˆ"GtS8Kݬ>"Ua^NzQiu!gk)sqsIH/ZƩJj[+DC*)IV:ѝCQ椄,0VطJ0e<,QڎMѵ(%|'ԴIE[RHHH I? ҋHcFPK|S;A$ԓŢKSmh5ST:8S׾"FnJ3H+]hd$.dWH! QStt مgAގ0H%G̸ D Uck\4Sk|FnxJtv (*6 :pg_wX 4shSs#`BI dr(j]v/ 2Fg-K751ѕUoYy<@Qc1fGܤ ̈́#Bltb)bDJp`Rp 3k'D1nEQr ֑tGdUmj(hL{5uFC0GRF5ͤ-kGRP6Z'o!@OebE ]Vkx5=)(Xrm+%RZ 1r 9 zWB}M6y-d Ք&dY\d`{4 {xWZE5p4d3lukwᗢH3fD$' G1!D B /ڄ{aV9 Xjt8VӦ4dYt4ҕjIQXe ,Äc*(OpHv9k=ȗXP p΀4D. i Q Y0⸱Ti /!!d&!Fu*SAʀv04K p[O hg]+y 9.z-| bhwXC-D;)f rogcyCP @A&e#kvpqAsڏl: EDiv%&@?xP*ZE 78*8\Ja, °" wMAFt26Љ6k L;4C:iˢ9a$X2'IB b8%mCɵ+z([0nx @p4D7gSMhĪ0]"_%d,<¬*H2bN KOp$,X m\Z];S9w 0@57[h#E_@wZ9^(|^D>hJMeoZ?y% @_LhJ V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%1~FJ b|6J 27E kէM R7H%FKV@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+\%9)F|>J kgWJ"@keX J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@r@n^%m5Z͆w/ޥZﻟ"d g0:[H)?+4E\!6zzt:]e]3r\/.^ +b#y8/{h+L6xn>OP{v7 u3-߾Y?/{?￟l74P#ӣt|XGϾ>:9zH :nv=uJB1!;tBE9b{] D0xYۍ+*LQ?8nnscmWa#|g4~>AvsQ ?%2jiiBk:#'QN#hqR~:fZdd1kQt۞|!~cM|7HL0>W&a@x0R[^D}hh:{z7=ζUiu1jw]z]>~:"1}<%Ÿ@z l4ׅoK7]G,lmi=ZZge3bGkB9S`"Z2dH8R>!/xPV%=ŽGܜ܂~9`0mOaJR%*Lq ^Q[$^=ZبqfeR_ᵰj48fB7'5~R L*Ind''w4'r)5h`DLHE' |攠%_dlp* p5XB\=[ɢY@ !$V,-"Rz)E, $.Տ3_*uZ13ݜyuintyzqyVu M0xK]w U~ͅSژq%sx 8ڍE,g:1mıSD`H覭OG!Lvl:lRc[gIMuLh3=H!t!OapBQ鬷(xxHYT>xA9yS& >E(UC7Á?lϴ X^),`oqPi% .0K^00@:T.OsRwU=~[/lV1wdydBK4p> xUbscf? 64䄙Ms7_*^,~|ra0׵^Oi&.{@Ux˚!yt<0rh>B,8qe[^u2J4Z͍b=_. 'jY5;˃/)SclȀKA)BxTL;a~E΅ٹW7-eW0oW XI,Fjja\ ; e)*gVa'frp"*&l~`6Gu~C[?Yl Jf1:_$1I ǒ oF'sK17=BU8yPg `z^.l8:K.[vSN o]!Vӿg&O NVfjWza(Gi%;p#sr#w1?O[ۨ7yw7HuE?hZ>D$3%&"ך:6~iټ77sЧz-%DQ(RA*tuw["%(#ID Y R/K$G,5g>MٓoSX夵c<$ 1Ӓ,'-Z@ٺ9Iс;UcdT$IpzPRVfȊsH  qˢ>G>6EbPtg#|ޑxXںZ[0/}1a,O&w/{6RX\'Ë6 lI/?Fgc1s]{G-rc38t 3u43߯o/rE^ f\=om %l1P A=%š^teEY5*iUZM%_GAyI\Ы#:呚EK [PYd;N\iڙ`%\+֜-Įq"'S+ ۛAX3XANIX>3p&Dz1=^Bx2Ms|5h?K(ڀ Tr{VdoM aI?w:"cV,fd "Si#qF1 FG/4'9Z>j'r飥 ~ q2_ & _2,$0$LG/UR2m#1Z`PR`KӋM&|ã_ej0zZGYe'G⟿XO̎S s[unnqxqשKi~7li`YMKX3~lD0\eWP{ ۽e[Q}0XF1e+P'c)q\p1Η%D"+,13 ?Ty[l4kwmUջ;oΆ.9OyƆol_or-n]u㕭 w BmVV1.,Ǻ >ޘl8+qV]8)U̪Il^7u9辽fs-ﻶK~םnvx<[_)ќ} v;8!Svee 7ĔG?7gؙVɼlА3ZGz(o BÓKM)3>۳Dm%*V`O7x;8ARH҉hILōO: 3Ii1&ؼ~3A׈z<dbK|_}z#_5%6}ztp>SoFU'i5LVP Lm5D!u!rCսPJ@ @0IZ I=':"$դZ 1veE2B2H(Խ]r}EyP|XH*pdZ%'mr&hR|ySij{n1;b١U!h1RѺ.X&'],D Ѥ}s񳬟9׌3e ECjLi’`B`IAL2r8 Ylg_l;h{4JY1<:.fZZ/HR9eh,hN(M\! z(# [/l[.=EC`.(R Lxm!V1Þ'<>f(F$i<ꑲ9,'djJb)#8%(9hw}(8jh1i͙o={*o3Lcgƌ2E™y+dDbωy,<٤hy{[l|J_lovl2]v>:6,h&{& wmI_Mp/^"N@bC3E2$[wg8H-iNW"(m8#)(Odq˭g;FHEJ~{s0VrWvG`w|z?r%gxK2Ԇҥ!) ']@>䉂MS\7)h&q 8+ZtC:V:O]#fwv34U=+  iF# I$CR[9XFDY<Džsa=76p*)cP IQ9/SRJgg5v)Wlvގw|OrŽ`v>牠0FG>gO8=?j<i֟a%qL"5!B,dlHuNG+3G@ T]g\΅4&Qˡmʼc|-"ur)\&4SKw"1Ax\zIȬn){(yu^gæoN/gFǻXe6T} K 75'FjD**Fi&ruPHY"!(Oa̡}m)P1u hjƤ Mל{BHF]& yخ*<&frG 䘐w? ~_hQ݂pC'3zy7<7]Q>XΟ^N6!375qc- V2$I%%Ll&&VRݏeV` '}56强+\AM Lv{#P3VË40U YQ-eH";1z9 ge}D VÜxQyD,i.ךS{7-V:mS΃A)}˶nǗ-KHr]_+[~\}e%g|_:+oyhm,]=٬YAvObArۻ 8Xvhb_9oxo=u>'/yz4˟]]]aX׋ RWWP*?D6Z➶nG-'<+(v)r满ܲ=P9+Ex6"i͂.2`8LE.Bm BFMKFYZ:Hv̈́RH#i(ŬvLg)<\Pr J'gMع6ݥY OL˵' >?* <>&5 d޵qϊ3{=/\;M뜁2r]'F+3ܰG_LP*{!·ʫ60y,/>};z+y^v9/8t |E<UR ~kJ{*qv7h/W7:M_~x{7o=9a/aa7mDl!{MS^ilkkXiS׳_]%hWP-]ClE\1IYKr-L2Z :2ci J814uTp vXaa<$pa +-FTp.I&Up߬<{b2,&"LyK7:vUН' iM fMd=POǣ~*c;W'Y (ךY +t f6"(Б={`%jM60ac ,9J 'A-k\%Z8R\!bHZD.j[; SR`)x.䟍ҁ 7hoJ.:*"cN'_$7'8Ft8 <8]OuΚ\.33S.FH,Il Ox~v<=q$?2+ lM2qLO#(4&$VG3Y k,z{T5JC̮m(Ky\*A(BOS3CD"Pg(Xp9̙;"eM ]Bt?KR#>*3MPA:%i*[$ {!Gt$̴S& :㤋I]ֲ 1(G"(-CbS8JJ'Sh_ڲ+%N GX79vj[uQuT}jqrK'zV%j!" > ɻ*"!`( jdsB  M}2fZsD M-3T=-%-THƱ`#ʕ Jn)`*Қ85c9RL[B( ^>.7xIideˉyÏd^ZV~h* 4ڌsL{zqWaձ+PmA,,^( G?>M /~:iXM {;ȉa aSlY-3 .fU9z`#`m%ga?v}SͶi3cB&1i O0vy}3ygPFOZ&=BW Rc{&z6 ](5C`2jTCwM?5/ Sʿ sYhwj3&]ѵzU[e|nRy7H*VoP55:QG៽Z8CH=)qivM@ r"QtjJӹt>pg;K&s3k f%!zPͬ}ܙ]JM:\[Mar?%&yH0e"H:RH5xbj'IgNE}sٚ, 0*cǣ-ͫ9®ރjTqhN7Dqm,qiOM{ab^ZX3 h57kJA^!dMn̨FmC\xރ t⊯)sA.cQ=_17$E3DęQxqz?wEC|m:k!ϥl 杝q6IivN^ƊRF.l%Z.P"ZڤLD $ $Q['%ls3[ M-8oQ0s}t{߃_7nr9x\9sN-~H6ǁ?;&[R)uS$\\O9V5x[z9aJEaJ̞M{ۏm qنjMeڒ/=Dz!d6qփE=tEezo*>G_CE{zwD$z62iqƐ2@e$"DTIHҵK{guf_nnQ1 e"*[^YK \DK)6" &Ğ ZN&)/R4 37LNy4.HA"JJ)__:8kJL}إ:+H"<_ziV]C[*z6{9*CagH%9 @I9bBb]oHWfn 6FЏXIrO)RlQeQöVwuWz8w4P)IlrxzsqpgVC4&ِQ!ptRLM;daElT}(%5ã1AK A*  "$"@]C;/ÝosxQ` FbF/4ek+8\29 = jדz=FECMىH <QesH*jjJb+#8%(9hwDpT7R"]i\L˜Q&H8x23o%s"u1 )7eQC\X깭;Җ+HC}0 ;"2wfd2ԃ4IQm[KQ~Bָ (I{D{,Aelc`ߣrFIQ5)@m H!/vt6AJZ}<nBJyށJ7k3 _xYA*/l_!SUzTO]Q=YтiA;B5WPyDTɨ$;_C8-RRƹt[$)}k*@=ƘgZi:rW܍nJC*ӳg)mx1Ý9ߝNiKWnH w|g}+]C}vLCk)?e~T_CeE,eYRv8޺A@})x{8HZZD\VyYZoSR$*ɭ5 @Qi DdZ:N|a%|QA\@P.LPo"W{Ӣ]9P؜2kq6>| Mo4-zc^\LZ`lN+qG9짅|XM̥ilW armf'|ԓ<(gql+ߥ?:sJZ2*^s(v =^u+\7?rݔGjHzS)#4Yd;N\iu[&,ғ+I:)P 6PV%Y2@r(dPwEVZ.oϗIcZ AXwT )MtbS ]1=^}"&^ʍy,UTON"Yß }՝] >)b콝qk|Mkнé#2j%PdR&H ].+6IX`7:z Taya!qj|xԖ 9DPp.۱9w8.z `CJO%jߓF/%"'jȽg'^HR4?n"qbr Zs&\1li4B %' ];AvFP!Iq@ˌw3_~H>=~kyeMvnQeR0#̾q!y;H4D+ d.l.j&V+N*F؃VKoMĔJ&VSi EUdF!qbL({t$/pWn]D M~_?8~>.^O]JKNHRǸ 8VEbu Z~8SzB~_>j6shy0ŁeUReXkƄg/ybJYE9;/sdwC.8<[nhVP; ~?uk%M|sV·M ! ^E+}ɉ#\__T[k|hcS2#4AMUSklV#vU3&iggsz"~Lٚ9p^Qnm_vͲc+4] .u=TfO@Q(͵<֎hrnoۜno6|[B, waff SyYw^႞+]kտIx!{ȯOd`%^"ϞO&/y}a::7'Çؒf ]么EgL䇪͍a[(͏+;\ox3m6d:|VoX놏L9'ieVwoeta=3z&fVuͩ#we|[Xuqӭi}Ÿš[+hro?xz_2)Ml <8(o BO3IѩفyoK#WܽsŇ DNDKAq㓦.-Kzj1&7Uc7B/P?"D퍽*=>O:ƅ/"7Wv>ȗIGZv…s`s]b\soe$i5LVPLa5D!u!r|}>GށKI*DТdN2A8퐵=!!QW,"O[`$A G?F7+/̽XfU۴Bk]Xysѯ{,<' ^ 65i-DyNɴexEΩ%EG`LxgXJEx gf85cu)LZs7oYOyj᷼Wr٢ST ,fɌk e^Emc.ف[:Iq%>%VPYr-iY5OړGԲ eIK US-?1,jc$B%-i2#|DxP3C"YC d m6 Käֈπ2F"NpT7Ռ\r$?S50DuQ[JӃ+ҿ ݖDz֣Cg.Zk Ҡ4MF@ԒsG"2#VZ蘿vȼȍJ"pkFz<:YF3R^;eP+kyOH2>n%A9Q*Jo s \z$jEL"mv+aRXBdN`VG ;"VqRUx_Gqab`u`* dT~ r דl{-n;%![a D P hgf#jPF(8fpGWG9 #٢6[)`=u\OR+@.ɍW Jb頬ʏ\}Y_.A * ;L\%HU•urr]̨R)PQةij[rjSK<+ A&`mNuy6Lŝ/Upv&}i4|{X[nv@[r^==wrqGP iIQ:R4E0}>fy9W2j']5 =sx1o2}fmJv-rf %nӐΑ4ȖczDN+? KSuˏ,doܫ﮲g,Yslve+u@p6ukʰ-SUS[LqX.nprL׿zo??_=󯿿Pfޜx,3>9\g/ |pk:u)jhжY øLt9q1F"V" tAb]ت#{IQ_]O%ʍ%s!h%sDhՍSZDpDOg684¥C4:x(u96>,jł%։ʔj>)t2܌^BX:{~x˲{lys<˗ʭ<7&}x6u˸ s޹NPн-qaddqk ؤTB !((uc4x_$j)` s,N!:RrUw:>E4zL䱫vˀz1"P9/u4qZ0+xk"&DHF\G#*rV\v7$ ).V7*%f];/Gqٻ6r$W2Y|?r\Q%Ex߯-ɲ,Y~бQT7,,~\¨c2^OТk|ɭn껺_GGRm aA ?_c:s\|=bX< j u:mF[tT9 ռ#ߒ5BEҸL"&ΧL %o4=z8j!~raUv(GY{Qf׃  'i=ҲV^_|pI8E Eld".N%hezŒ8Kᜥ{P]+$H2KSv,D\#%,i+{px2os0d"CpLJ`<ӔL| 0Ou jթLS1(OJQ-$ds( ԖY!JhX*l&&,7Gҷ6h8\~lrdPP3k|6=Mr*jScqZvy]=L}OJ] ) O&Ff3=gZloǮ`&gYN{Xl`.SG5];|h!GJ_w N8F2 oP ]Wѱm5Z({PVnjnhy#D6:,➶o[P|~0uspݘ 7@ל{-'Y8`ͧ@NODrNr;] (wYZ t;MrNek+ݛe`VpZjG&КZΟ-%i-i2A)M3Q-9ηy#w~^;_ey絒C}+vG|+tύ 8WcOfշ̮Lseѻk* Ul{yVim_i{6Ψ|eD7 ʆ6EɼF|wRf⼵DJ?OJZF0/=f_smA+j]V-{&i璍ޚ~186b+6iA6삩6:7+ZZO)7adf4]jp5˽c q!}fe?ldx~2ȷeN=) " F'%TpPHZ aad$ JEI#"J^"$K)u'o BڿH&嬬=.Y;%y)Ȣ,'ک@+-}>dm]f'qDa=2.VgʧcݠhJJJjcmfW:*L[@r<487Um*{;eSTB| R&=6&)W1;gr* 2%L.%V6igLIIժ3*|a5x/}}k.w$.q{~v9U/+@p8fW Y~PISY R{B@o:Be.K R SMƱdK&#e.HsL=vajVPh*E/f^ ʠE'!"*!SXldEQ ,(:~)jq(<> xElC я(tO²=|= >t qt0 ;o0%Il' $s!&@YvEE(7yʻ0֋hpC`sɃHZKaF,! [ |FJ̸H>9x+DuL]e\:w Z.+DvЃZђz Ǥ-kZW>0y 1i(H~VM-/8Q}e I3.e(dqW]J$k ,'jp3*=惘។wIy8AuM.=CsRcJ%8XV"폗,@&xmٟ9)ʉ ".yhFtt>iʸ K9XK>#iR+8 UeFȩ+ark!+ ZΝєQq os4Yq w6,jl#ΚV1iC9N I!ysqZ$ ?4dJO?f ܏ƓC3]èPf?~&`ҰyoZŸ~j7z\ÛR7o772Md<˵ lK왟]7^pQҟ k>$}u㫏ig^АF"2;YZva *,?إeg[1R֜ ¬ʹ9৒ Љ9s3 G]g޿k>Q5="DGdmæ_._gw5Ǯ2CWOfɚ)iAYò;jVc8 SpR~R"έNʸ =:QWG~60<͑dБŵ\9"h$;tes ]Y^q,y;] 2RSDayЌǟdȸnuo4J #/8VvY/|"_(D?lp|ZM1檯7S/%(ַcEj=Ge\~Eu!VSݤ7qwY~5ͩf!Pb!Ic8u(a Q$hJj8 24>s$;fCZ³th D̴3oYYe<^9 6g4 oiydBw,WkWƆl1ˍyfHrkqH>ČN@U0e?hE9<泉u*zA$98͑[=Y!3^i'L)SF~X{X kQۅVb7Fg3EX͎)#s/؋a5}`T T$e8)9)ҬD"p"(HzQb [H'$.KB#ǗstÍ3Xp)0,iB)ƅP> eB\ 1eRid9km,Z)2Q:%zhrCn鞑O ҕsPO4N&!3ALky)[%_*OJ `X- {faOTg yXq3S3Ub\h4Qeٺ5Ǩsҧ`"V32 SDK⴬q|l2Fpe*.L3) ) d>FXժaO_ }q d,^du£sw]"Ff?:p0S hSNi1;#h2Lp&hmURhM׋YBht ^KksTJheE3h2:TD"6S%% Tp zdڞ<)]~vl02#]28HbzQ1A$y 6FL8gSe|c6FHkIꬄU1or2zN(e@+S6%TꅖB,O>)usycB\";Y@ Up%GD TRk[-稹eNVZr9﷾l}jsnoa9KX[z #{F>{'0'X{47#՘Z|*@)PjLrл -,fy0k8-E46+ٺeBﻯZB]nݷ+f3X85MզKvPΧYt8|[V]n ]8>jMfU7HK+HhN~b19p"M%I{Kɒ򾓟 Uϊq!D ;MOs*~'_y6²ًEbU2G9IɣLPכcLNqEK;KxS;kx U/ *`Ɍ`:Tp8!v^{ zv*z} >ƫ/f**z}WQvSӽ4IcR#g+m# [dO>#,xe1H$Ednؒ_W!ϖ~xrT!fzJs/_ K#} PIk pZ)CZfڎ)QD@NJx4$hscvh5!5eav3C>Oq u۲hj\Zw):0}_ϴfq ho^ )c) Az竭e:,аʞv6C{ Գg˧Jr ~/QJZUsX"Kh$`nqBO5ĝHIwUzˢJ4,zljK62-}B;S4:&,c:S>RlJ3.> 'dPXIcs@b -x;7Naq>@gWҜ4a_ v@̀W^F+R!@!AhIA+jʼ:"cVf "SsvqF1 C ЄyϑZ [["?Wݸq٭yZ_.Gdd!$kx4xHKZ+%6 %(&q1dϢ2/-95(A1bHX=))RR)r8xO h(Ӡ EdO} i'A.ӏKDjYl8b4m~Exjo꿰QTŐ<$"ׁJJSME$pJO- uR`|!3[>zUϠy~4#S*VZMMnYİ ( 5ޗ`]6%%ϕq`X$, ʱYΐPUrNZӲPǃ\/g3qFu4SV0\t޾~j4Q0ը-Hih.ZYM7y8oo.Z'սHAV^7 &¥47i4sWn7WMb}ڸ:/RPr ;~\AcY|PئFW1G !.GOV݈F,=bvKvDLGs<nHi]F.zz O"|IIN.݆oqk/7B@tgǛ.q71s<4fA@Š/NPIe~Yi3)+XCZ eHT~C#{װ%ƫ:uxME|u j{pr`NgV m n"׳~x`'̗) *@{vPW\a1Ti1M&>37ח ar_z'-\+rRe֕%rfD#QH0P!wCl$\ nH㈅[[,zC'.<5u0CT3XNet 1\kv&+w.TEFSAt{m , a׳{|K&*_ qhUe{ 2j~oZ_@Wo:.R \5^"m>__,+ :_j+d7DǾ޴?;Q\gOR3o`Գ0Ktvq3-cEfrϜcI;/7=v3fiNj Ԏ*I{&10mg6zٰYSԉFm{zGN٭LDaK%pyi )'eASI~H [CSY3[C-YpȞiy(*`V&D.̈leni<'\#w2mesAkIDGJpT&o ,"HZ"T i㩞݄<.hh ?tJ9** T- s=A A!Drz0bQ1ٴ 49!b֡'(nc8!=1:Fm8~^;ƥ4u> ڞ~-:O_ eE"V1J-6Zy Q$ TJO_, [LrOa1 ܪtrm4 )h"b"hҌh9uB,+g$5LYB=S$eXF FC׀rV嬖1낫v0jtD T9k7GGc2YAK A*  )IEAJSozbKO<0$EД Q>9 *j2IӎEmNZ$䢯9]$O)C$5 8bqse2Υ :$Iu&AUI !#zj1!15aZiSbsw)O#|v*v'_,55yAEP<c~K<ˮT9|VuhY@\6s!.*G@/00!􀀞3 w>d(AFZ_uLrh) ?s7C/Āz1"|4670&"Qy " >)7=vfZM}om4œXxo7}h㲙!BAt6m_\kA~F߫oͯ٧iώS"ܻ rr(O^rȣ>ڔN&E&]5mI5m057FHH.uZ7<}';SDG)DW"PJyWSPpjq [(]T2^^`>Y`)&?2@)U!e&jN2ɲ5P2o } {KDy%,wTM-VH I[ baLIB<1yPHbԃ =On3Bl@{?rA]x?_~QBiA].P[%j 5ڡ.%3}6F8ekue,TO"J*([ Yw<]h) ȌaQS#*h9LuiqQ5KHe7HIȅƅ[ss.s9#M9Q[ʁDQ:ɠ9]kSʘ?m]ײ5A󨼷 %@Ai%DFfI!%ג&"!'FLM1|@ocסiU L+`iJAhô:iCXMFhhqǦk[vz*^!:oSiiDNp !6n`ogj{7_!8AvTa>fg7|d6/նβxWMI~lM d,4ج~QeYe!93JmDz6.Mv&&sRxh9R Ш4Z!4KJQs^B2$1ȍL2Bbwb5wi5=ëp:yA_JDqV2BqIo"|<.ch&>oJikүFӸ<\2|[⁑ē 35R; ԔOS .DJP θ[M!&eM'ax ՋsV3a+]>2Vvii|5³6BwķdKԳo߮Vfov2^ Ra7>*jryU&]<<∆\BVsWמiqu5ku~ŁwWma@({)̉4h޷] Yݑ]tGo5-ʕ-Zɺfff*2Ąqʔb{</;lsp2pֶ*ݭ.r]k)qrǥXn}ڐU]^7՗7ma ^xRW?~~6hb)zƂKҗW((Ui'xz2/_6h*ZGդ?`4?]䈌d緟>ޗ?|p\؃wXCe,^'•wX~}'pMsU޲ii!M.zhW@ݿhWqE$.xl5uq'G&>'Fw'3VVJlɖ91 @f S9/,<K粅 /8mw&Cs S%Y4ZDتl 6LlKp, BZh7 w:٪L1;>ҍ.vV~, #X6U@z6 sKz\2,P YQPGkWxo22@7Asܛ{םllʒ J{HBJ dOny@lhC%..κRbȵLV\X)XJ>WKy. ` 59") Ngp&]0 sV_!ڽl˝gϜ 5'kd=e:;/GG/1 |Tv3dxz85 tiLGˢF!K&uPN֠F =*J5GmHY ʙ,< pLV,uքP9rnk%ɚNeo%1*דg) ۜSJoh0YYsEqu/p8>X~8,2]H kR2ZUP^eQm5.+3P7`2!EqDcٳ L&nǬ9bfخqu%;G ݙvl׆k.2s2u5RhJfng= I:~j+o$w``bȄ 9 ,jFM_")"j4jӝs?lI}*:$,q[D;a:`4K26@F!Ou(Ch6NX/7Mz| :lM_{\P"Ȥ5HÀYr,qOiA'"mb:2 f.eW^LmI0 \)Wx-k͎'wKl&G& 9Q2d'PYPFI:M]jOy5hA. e(dRm >Jm%_T;'C?ڀK_M: 5Ay2 jpk M4f3e|$_#ԯ5BZMVeν)B2!{zN*e`D[jRAxaq%1>9{WYڀj߼̟Wm|&@Ɣ ;lŶ L#QkːcLV'3YkQpɘO|hɥRTs>+iLo>ENn`Fqm-XsCT4vȲ:Z͑e+=R :Hl {,# HuQujM@q<30NE-d0,͕w1cge癛OƂrL݀r8݆3E5]T_hi+0y$ݖRt+`HJY LBD3\aG zFwg:ktke}']/DDY9J5;]@5w5j5yUz((Y DD t0ƩTJRPLND?w K\9".y,2#::'|e\eD-h%Uþy`S&ȩ$/DCVB8)oDF%$:f2#e2αdj(&z0 [ɓ$ 9jtJ`ePXPG7MS?h3a~㟧.} nn| m;B $֤w a#;;4ya[rnv2.$cN䨄M'aẍdՋsB`Dp =}9fEV NW…luiz4& Fxbc<',f[~׷oWܾy +YU",D؍O V7)SW:%$_1"B>s!_~!ѪV՚廞o5%Ք![#vvDHMϑ? h_44"|V{j4,OߟMq}_s,gq,)aACcd:D1$Xf 4ٳ9`,7J72TV(cGB$[6HP'$gf8dωQyW;ۣ6Wn)"FHr˃1GϽQFe5,zH3t"Ϥ@>K +EkZtD敱!r}` SVCLXĬ&{JԽ]R>ڮZŨsme)䳉u*z rN b{EVB&yDʄƶƦDL >[W$?/m]̋^fji%l0, X/ƏRxяtv_p x<1iGx)|K ) t2Oz.n'Ozf͐Q6!.eQZ$ @ӱ>|ʙW9!Oɉ(w8(mങǬj#?U}5dBsۍH4x5< = ͬ !6W)IۚžQ}")vi$ʌ#3ڴMdSb+ch|nC߮^ߞZ;/ܜvXKl]f.Ζ{ ,b5}C<,>n>b4oMcOfU1<7}gGz+fuޚmQddw!|6ys(;?_=3>v1hѳ)!"RY'ឝg l{3&94i`C9":[zhʾwzY %P6v덅/x#BAg En^.?;Z͢< U|?Ku0{ ˫v{>~xAZ!?zrjom%QXلklw=qj'ѓ(0Yhb[y^UP7>a3gW jmdvpP*mԱ-Agh_wO-%] nK!^6m7wNwX;nm.eidu7I7_]ACߜL_8P9|S{t )q^b^׎3Ym^f~+nzꦽqCV}鋓nBQŸ@ Yy{\hqy:üjUC1|}f;:sh-qLɱ%eB%ilMk,]&)/rY;Y.$LdZw{WSNv S+g4îk/O'K$/.Z9ѺOːYmm5ɩd"j4-a"d>hgV&a6*ad1XLǃjQ唍 MhB"UF{#}ʹR"(Uԅ^TC ٜjliMZ)nnOw΍z io-UAdlt1M*FKݫL=SܤhZ٨͇:r90K[[jVN O &K3mDۤIM#$\KSmLb0RN›1-.mt.4B%$kpGK8F}x"[eY=ZǬJ{wa( oR!TR2玮B;L.\:T{}2-8[_:GBF`.Fhcy`~.^7,HUƝG^%cNօM,LOQ'Ss.?nN iTE%jgwoD $YۢDwAGo0VطJ0g<,QڎMѵ(%|'ԴIE[RHHHI ҋHSFVh4v7$ԳŢKSmh5ST:8S׾"FnJ3HT'.hY[V RcI׆RDȗBx926XzNc-OXUAEE'N@kah:^vNi2Qh?(Q _.JM܋Fg-K751ѕ^!B ;e!U3ڄ`޲t=/ ((Ф td..h]icm\g 17ƢPaVB]$1dlDH!(rĪMPk>Zl?c%͐ڲhF& %tD[vjVECU}XHhf-w$a:J؀ `F}+%LalmRj6m73gL Ybm[ٸug#賣]tܸT9?[|!lL/lO3#S~葓jjφ;uM=I?a픟[y5,ϯ.J,-B./ޜWA/rkKj!pYp*F'o?{k>g_Nyh^kTmVe͗!,Q05!δ3G+{zcօ\˴={zTm]혔=i׻i&TgmIaCl.;{.dA?-]dJ+I aq})a5a-h 0,W q< Ӻmz&l<}3wk!sw:böB,Әj(ci"jf,{}xkҁo庇BIkkI)5̰zzaMQv{3[ˣr/mKu0|Y7+ȩA~f֌ ns; K/8$WWbD,S* cjV ֓P33JQ:}r8Z|-Wo 4nH5$pbзj$XEIiRwGb$Or I -Mfkt+/" wk@7u_R^[yi_8S ,hAz\K_X5x=n<}Egx[_@=z_:I%d|Pg4/|Q^_Um6w~~|.UDjݚo"ploFD{F&ݝ`+[<́pRk˾MrWO S@<1 F%]k}whVVlt3] d:pzwA=òjZFeHǁΠ?xTDUuI=-TfN^s˚Ÿa,8Zp[qx}7 g)B< %ELT#a !fcJ /ވŠ^<2epbZ"Q>Sq`L~uɘ1XUffM1F3O`_*⩼B|Qb\k"Q I&pxJqFS92op^ Q5#g0:dzrۥb=E=zEY>fI1\ȮYSU :qlCQQSvXkzOޭӨXs֨3JXFu",}hpŚpgl$8i N>FRB43JjUDQ5 "bg#.Iݚ/Ȉ !(mcpT:5== KS'qnDۏ"-޹^xҦxK.A+ 32Ia"m0*"j<Սe#%l{vQpGgZ(WJ% ESf`s4<EQ8Jڳ`Kg9glIg5] ;h/@Kǐ UJQo*I-Cqa')JՊbQBG9I9?{7> `b@}g0u!F]PdzUY&1aVjF7Hh@Dl:=ӓ{Σ @I΢-lFև1fҹ>0@9F58i4N4i~貿G%ciXLD*(a3C#e)bk17&^)XQ[>v; ER]v.9pf !I83 \8#W&z [jq;K0Qڀ#y,'wB0T |F*mΌ.An>ІqiKJ-H$CBlL&Q,Yת 3u ӂJ>%7d *d k,)PV8dBv8.q3 ('G,7JIF$I3' N+W@yd(g}T"`M_C*#(e WcB)7j<[֑aYqGkZ ј;258m.tN4uJ{'}fnS{܎[ vvoLWX'}`;ƣHT_*敻 -c1k^\gx Lq Vb҅VtE2Nz4 1JuJA|BcfSx \ 3f4j|sWL#,8ZI0{M{hDT-AX„O#_syzr>BٛAQ~=Dtm> \>xvۢ}9X؜컒bܗ96Wpcb{ʘi%i"YƖ(fYZ;~DgRej |zXͦwu™}=0`Q{A $a.3$M:A溬]VsʁNJׁڙZx]T[LN6H≷"'ۈ4*OGC4(‭iU"bK(F[ :9A!e$l1U4k Dn& |n|*I4}l,_ 33fׅN^}&yU!I.-%LiCp +!(ՓH L)^cjH |O(҉>"ҝwyȓ @yDaR"u>_2`.BFPp$*řŌ:Rb&y$,V:3{򔕀DX/5<+ #s-96x|}Z6z4<}݁ ;!qT,x)513Ω8"|\j"e_uPm*H/OW)StL*"Q* TTJb,"B:PdwYPzfI7 "`XWA 8^kKjm $jϷکPH֑=M0}>: ty9yx^:rtXj5f879ZG%hI:WɊTɒ4>O]:U~^_ M;;Nk1ҫϷ^ʥ\wPL+^2d6#MeZH_t&3U|r~熓ɟͻ_7/oz70Q^׫wo~z {̂KTm SAc?{C\Cxb sκY#%,[ָ+_2i@J|Kث#mJғؓ]ODLQ+GM^%##$c@u[-1DaU'}iÂaeNH ך3iE)3 # MɠmL3k Bzv\sJH 'D|&vvUhzB?4%,UXZnm60+f(fp^r-" `BK 1$)n;˜o*9&"-34dW'euRֱiZЮ.h ]m'i=E9W7#}9Uz2 N! Ɣ#Y0/]t EL-BG։S]z|bN%z- DSi>jr:(s޵6r$2X _K~(X\,',Q-1HCI ߯zЋ(j(Q^jլjWE%e4jAv7U`HA2R!8&%0iJd&rb>Gb-8[gb6AvUh!!skEid 5 9 ` DZtd1u c˗QzLi9z7(5O.`Zpzz8dz7*8'(ţ])~T&*+W/Gc)՞Р)4ibtʚiFM>6^ ")?*Z}F05Tג<\ -p~@ks`8O3^C3 Pfx9m_p™ c%٣X @Ԁ?|}ۈћK/Q =#Fj@@R ΋ @,i.'v2%L9ݜCS<+i0f1ol:Ė[nOw^D&|z}k]Hz̛<:SoWx:5N_杧^3Pϛn[lCyROA=]W(4e:LF+mJ fw DФH:]8o-*ϓVw q/k:G<Վ $ %59-<$SsTDEv3d &ڴA1·ͱ+sw q/'W?']_Wy<99vQ tqH7yZ&_OGl/md.s\,۝ RqN`P4?<>eݓ0m)a7[4OADeN@%rx.vYINʑivT?+V>%eR{W.Pj+*cm3yPtz0~ۇӭVo~_m=^Pwul^)*!>IX R&=6&)W1;ӪgrPM]rI<\YI3-ck< 'tvKRlVl\:{-W, j~wzA2 5-b 5zgb1xTZIMvۊod;i-S+6yb~ ^k{={o|72wk>6=jqP]2NHTRe:Fm];-;R8 7-=f]1t.yIk)ld!"WΌ;ͽH3zWm }+NxA3]Z?~y<|~.!lC!=3BV.8?{!.E TAL)3~*(Po5u )R/r˺HV+_֥ 1oni7:Z8 .NtkOM+,, c݌t+l>6`"FRy y1- Q!֗(ҸcJm)IJm)VmS+4㙁2.j.u@,g%x. "f0h<&h3w37/x޶Eo{˼rPU#]oDI]ݿ]:Zq%e|4HIfqxѫ'8C0Nyûw ϑJq**BE.%R g(L/쒗FjAG Tc4'kHԊ'<$Z}Q!T3^#YYprl pdFxKEɊP+1[habwlE|}Vz@oeRH ɟ+26៾i߾p<>V e׿vk?WiX}Q4!]uǃou]uHrpP̯W2f `)P6% Q˭g!3>8헔[UAx٘JZ>^G yizO .Hp eB9 b:ˬTsX2) %R7ڴ&=4# Z[^{C R,=6ļ†Ti݆ࣃo1B<&8'KO7pTL4V?? VhrN,6Y:- /'~WPӰ ׅF?tX}Mp[{hfwsSϟPQ4q3„|M=fPD,$񨟟\?<~.v\(ЗuOZ Ve=8&l>FӌNͤ|3W V\ݝFlɉQٝ4gKUjzdAIKW=/ӏ4wr월&m.,N8:NBR)%]I_~F#_J >;\#A R(#\HO-5DOw/ u(C 'RDpn$Z˽ǥW/d r+q.FYeVyFLd*A`dݚ=سYpw!>i|ʤT9DGOn8*!^ ]]Y=:旦Ճק6U1 v^&o^4f,=gP癥C&QjsZvd볆4Y ލ;zéEr&ZJRe_b0MawWA&6,4U$)-2#%+K)z ޕ$ٿR wIyZîgf^ۍmisL4IY7xQ6eS dV1++3"##gճz7𳻆jX1 N3+;üVrnn ￁颏IJVy-k+,GjAsڏj}QӡFid.)pu1Vmi<A<,d}j)T.6Ġ^nxe>Zix2Ҩ/1䀖IՄaMn{WDnK{Z˖T\y$Q|938g7;Wh'7wzR+5{KV_lgS!>E)+ _Znluc7\niZ(1rKIŇͣ]/LƝ?@vm6_yK QG97 W5f5"tω5<"4ݦ*'5wluQܟWj"&*w2W4AzA FQ)O1_^^^UKW6e-DbIFL]ٶ+<]c({:^)uPJV(bx:~5"Z2ʊp46M0Ƈ}&q`&{㦍ty-0s!k58`mZ]n^͋ ^dx.mbHz"k! Cù&9g #߼*t;C\)oةvt.Z|^_l9B'N6C3"a+Mrʆ K$# :C6MUp:WyCmcE+jfVVh8jf@Q6PJë1O´yU=[+NQ~tzڱMn['F1DEenu|#@!kUĚ\EٽGpƮOgVٌ$T4ْQg>e O%)at$K--|bğLFfuCSuJ dvڸ4`\Xm +I~PKEJq*QQ(2ڃqy);w\Sq*{ifkzk -2w=0MчzːVd3$u,Zw> |``WUE |jQJOuk)aPY/k!.峾w3P;taN,n15w9Sm>YA-fW+nps GQWL6׵"~ABxM }EGTEk;2%0t JFhQ!"Nxpo#P[0j 1P =I{M$F)R!JS3U0NTc{WWvNNTTp&%e%7i-E  "ĝTECbʔevhˈUHd241J^AkML(Juga°k:=R?:}YYӤIQ9׃.LCxx7w6lZWo3#29nBMNh)sh&#mΏC|5AZ2ρ!2&Q ZdRI)RBS2*S{ʘOfhz:&YtJ;UQ*KhJ[bmabdk Ea[Opޣn㛌/wirGl(LrCm[제DD GҔ#uJ('* j^:.EXR'-C&{.2½ge8(ᕈ٨r;blb‚*H$hSRm}(Xbq,V[ nxEU R3ʙV[㈏܊~dS|V"Nsdm-`Tod>N"2YtD A1`&(hc$ՠ+lug=lH}>x؉ XbE-Dunz UzUzzrUOwo6;Yt{ޡr0Z3NÏh/?\JB\;QtAD퓓;P^d40"&mR0S<ŽKߗ,djjw qCb_nRH$Z >$O"' MQ捘r9 -TDhiE1Wla֜ITꈧS) ΒVXugQW&yY;0elJqvzR7zҪ{xs-' Īo.C6 m+Fo(A+5YgN!mEL[ w6Ñ:qE1=w@R=bBC^+i x') d,?N(*o $rr♇M=|CWhnv@.lxS7K7}- Z.W1yK ~d $xµ>jʅFbwR" 8hr?~g~k7G ۄO .)_T0*KҭW+>T.?{zu笊]`r]nmⅨJjN[(m-FJ[8ںW_ jWY D(1A2BB#2IƬ AA2^sg`$2)[c}Ly/)Fu7]z7zW@g)Tuoe{U?&?gl4꿏]?}dzj4ˊ":ϟ?}zUW.+/ϟu|z_uyyuf|PVn'-91^Z\uY;-Elrrr-;M~]ܹk ϝşL FNP_f8${ )gMЪNd% 60zn.UkGUb~0Ū)4A' ޡ{hZ-k%0l-o'p޺^W]T"{!cVW.|L u] >}dj-eaI̡iT Tֶ7-N7Nj-O8Ԡq˨P !b@.ʄ 7]/k|Υ CKf:I!I4w9ƹxw<ޡuŸ3 =EьH3>G){C, h~ח-qpuJK83vvֳ%$b/%\ON8@TX/?ejyjfHMj1QgJ~(7ԫB^ˏHazf?"?zͩTG2oA$*уqBjfw{$9d3mMFQT Gu U]["%(#IHge$O6TK(n$. ZS@# &763s *3MEmiJ/֝{cZ ߹$,J E_%0%c:-9zeu$8iQ2D@P.M %  Z5Lb$*qRI.P3od&GC$KAtIFrW/֫F0|IǎIdT(ĥ31 >#CXC߻hP  xB$y4\hG)#>'&Nre9=>SRe ,̊Hx@L"p:)Oju:˦brl ݂34;gNV<0{fB; E wr uNᩭ1$G"NUNUNU XE($u6 \!D`Pࣇ1)uwMy|ueJaRQ4+c>7}0^(W{ehBHT&Ge;A! "Kh,i.z-mJ(n<(':LȻs<^7xhVZkV (A%VxY%MH!Ru$$0ItigNAoOD)jV3LM.'ZFVgcGoa"d:xLynKvh`e{HCA6Zhr?3\hi<'Y?-OYjZRNYt[wVTYp 5dATkdP!S|q]j:wsFx:R7_ӱ=n+z4K.$=1  ˔ /ID LJ|}\Cp"˳t>1vPkKޝ#@R*+S~F*\Iv7neBI%T`=o (lKcp4]CΦs7D2bUx'*x zQę|EEY>/7w/KA`"XP+<VI-' bI;$_D)Y93]|zKB!˵ѹ9:Z%$ $4#Z#yNx~vg$5LYB=S$eXF ƃ'g5_}|lVtC wr$o d2 + TN) +*\YKxH/A~dO )r)~p&(R E< CCΠV {=ד( PI9; P*C$ryhPlEt%9'*_M| n1I3$ gOf歐iQb<'"P}[5D˅Ϋۺ.PfG^ա0_g$/юoTh-h&qI&C=Hl.q%U$h\q 4%gc 0\ wtF!r;s4^V` nkmui@",zgGr6Fvoad q_`ZHutZLvCqėyV ۇ.|Mpzlc)* i#4XxGA/QIwP4Nq.#4|onJCr%m]xY!eN6:e}֍O.sv!CG>* qPMS(TNQi"!myihH&@@hJ*E 2`^נhj?TY{ '`vV\C!V:iW]2-cPګ?v=N$@ --VyY (" ODH<a U[k4O6SӚrty_s9Na"[.p<"&X6nr-7-m~~o~F-6+GPrP"΂*I0ӃfE8qFOThg: T謏 } F9?V29#AVΦsկO v1YVsjv0:햾ϛ=igx:xdxu~T82,2+hؿB ZBP沭J,HV~pk .axft9uD@j "Ss\&AQL M6ݧkՇ!!=1Nׄ|Ʀ3{eΆK#Paf5<HKZ+%6 %(&1"2-Z^0<''HɱD-{R4RTIRxO ?h(5/}n4|$INK0.^h"pCp4wh6lCˌj 0u'Y!A!r4\`!$\Li9`UzCT/ӑOggC}ɌBŘPHb/y:<0tz%AGGq?1QO/`Zd#e`VTdݰ>^8HhV]w4Ͳ?9#xT4.>KMo?%ONdzq߾Ex,MOaVD _Rx5,$3{gZ}|1:=\moKvHUиwTRHY/k_ڼ##_Z}Ȁ?Ǔ>O, &KCؖfP_oOMy9̪{$hTUW-3jTҕwF}.`k<=O hmeVw76̻- ݘ1jN&*]mh7]|1tknYQ$p}-8-N48}٥2oJf!A#gF+lL[1>B9PMNN Ձy9\9 QJi"I'%)27>ir1筶),cuЛzC:-BgWO_DwǸe7,}CQҷI}DL3,h#2ێ hNR e)DdOrB1gϹYC d mKq a$"wϹuYeGm) EUΎ~KizH-ȌjwmY]E Xi2hɹ#!I!%ג&"!ZOȍJ"p2G1BAg)}V4) " N$chJ5r8dnZUÃܗNag}QYw!Z\ˀ$ut*8l!ʱ 9:)G #ڲk ]]9̶-ΊQ>bmJv-rlr%i]gq*%%=xnة8u?_j6B~fzx:Rۂ ZB6Q,+M@Cdoh$] ۲T|r3d?ͻ_7oQf޽wo~|g mSXN_3EWL݇q܈]]e%Íq%5, !1~@#gpyCت#/ғĽ'yFd 9|4i)-D"3޵u#ٿ͵(  6v/YmIN4{n%e:"Ht.ŪsHIΓ~]pȥv ʑFi\JRR%/}U֕d3>}_:b>F7~u!cww^uGwU9ZO'|n1,e7m#=8#sJCSU E34dU45(_Ǹz.Аpj)FYP eө:R[N!+|%Sn逊b_+W/V=u}ʯӽ4CHh9 Xbh,ĠN`KX,W\'q:t(^l?vt]7/ t'ێ>2RXi8o?k8=]N # 7PqyաS fu8/Z:)Q,We *Ԫ)E"VZG>˔E֔\H6Lj-ZekX5Mje$:S JD)5[t̳.)b,pY\Uז}-`^)kHYVT}&i䂾";7! >>a5aE9[\mޜm.AqM8prazeOXr#H2DQliVZ)kynOv3i}VXmn}?ȁ>uv0WQf:}7~|?(<ㄟ9-|t6 ;bH7!ӱG;+F#} 媾)nUHx^u{We{F;ߎ8 PϞOi˞iYC*05-Ox*|帲>sl\7ϟ0%ga}/^9O\4xr]w>l|QL]~lc_N\ixm}`.Z'UW߬WgoW# /׫oU#n}RƐԍ =N|;v{ l}\wmKOG'cˬOG46L7lR jU8|/Vg{#Hms!ۍaCL/a_;Hֽݾ~jr{w5u s99Qɓ?z5aRęy׼^ nd;SO7Gҳ<>_}s OY+mO;wCOdL%ajMMC" ɤA7*eRUQF!j8[6],C&AU'2M[ér|n8+?2qFau0]zSxD{ێ.<^TtE|?K6n.d f "˷|bSfRc~WO|bo˧Ydz6XG\x $Ǐ]ܸa;nz;m˾bwCkx7V1S#_{J N! }nb>5R9* YK8gXI|},pd{2wc? C͏a@^X>^^h&gެ:XKrӮD+d&cߏOx̯*hw?% V?^Bb6<b##+$V 3_G H"Y@-r,i4IZ͋Y=⢗}HJ4ap1(liWB37F\r^$TNc :`ўB6=Vɥ\.\y; TJD|.ˀVA @KK.!Ìip\,5e:d:Prt9||uOaV30-ENpE A-)J# I VWe2'Ea8Ӆyc9p #"Arѡ^F=ЭH!3X`GW}ƂչE,TGWgjbQI9m,9xY'O7VCteؖp`y\E4Χ,}'XR Xk#p`WHcޣ.)6H/9 HT`%\Ha("Ft3 S@ʀv1\4KpQƎn<,L>^|6ٕ;g,1h,MD0jik 8^ A8 pdI..x]ycm\g 0S"I/&CdbzcUjQc<_ydQ>& sh)dl9MJ6 L{BhϢ;+c(X#i%rd-A(RI[oETĽ"nZB[V@"tAT | ~M{-9KPa7Lpj&4o8hk|e7Y|v4&I0nd*Zf&)ԍC7@7ӌJD3 5pSQa-~)(r2HuTtk̊ZRAΣDjK7fSGhV. 3B*y 'r ~D6@mJ e0CkvzI,c5:#S :( 0rHAO!HHY@qo٬dp4 *Oc[WT!& <*@jga~{"w^^0`\SLCӢ `TɀG1 f>9X\GF$ _3D:csjg6K)uM@pEPAR䳗Iu2 ha\5gm} Z|Z-]4%S#ob P[;i* X0m@…E'X)*:LZ8bA+}!1&`JnÄ g9>X2'EB=b8%CI F­:^@\0 u{cXb*5K}`1>옄CĪʔ3#I!:·.P,tR@Ǖ1[`4bˠ2ˏp]K!wTu)XoL0@5G 'm}}7{5^q^CT|ޅ0cx <X}1s/Ui+lfsqGVR:sPkqQU$%R{/G D(r@瘺Ϩd:<+++++++++++++++++++++++J KB;=),/Ž%gϨrL+++++++++++++++++++++++ϫHB/N ۃ/G Am͋Qy?K (<ҟJ?(+++++++knVГfn_Ӵ㌇I[-J.@=h=B\Ac"v]ndE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&PdE&Pd}L m$$&4p@\ƂaAk=Z)"sdLLLLLLLLLLLLLLLLLLLLLLL|@Gd.A5Rep@۔e1n4?k&|xcG# f٨vMVgFܒSF̝dQ~kbo=Of{{w ?]j-4|hfiQz N$G`b]@š+quiޥr~(`X$ 䉄1l|S.`QZM ̀UN\uaJԔ)(v;:A. [0ExE};=_ f 0@8hLyK~.H]&|ږ-9!˵t!_أeOTҤ[b3%ˋdˌ>tkmMyK'KIYKPk'4'; ;X-q0>>q$e}|ϞYꇬ+z^F*l9Hg,RT_Y/֒^}9FM:烬8!>Ò}9TxߑjD+ã[wK˃if_/~]Ҍ©Ama3s uϮ/_~[sc]sxE^\ޑ[޸gc{;9o$gv_~~M ޏwqU?ܟwl*owƋ~n4mڶI*շjiLXd3=P9T+R )rL,Y+Kaxa֗m1cmVY~RiK)61.phaܦ46~\tɩwd}tyq`Mm~B| VIV=07 0S&:Ƚwʵ K{`.}! ށ hC|uhݓɔ&A8 ~d4?R^O&ᚃ8{Ն7y&~" Ĥ6 #5E[mV=L䠎 S *e7)΋Yj~mS'ڴhOS"vL:)y~@8,;hОQZODB I(4 zaZ9T@+C?Z#RpλhX"q53cS}p'uM4^oq$QT! yШeOx&%0qAηy/^ҵ Lw)q5D_#0[V㙘L8":53w:.%,]+%Ͼu\֒2K zg8Ҫw<3fzt%Ŵ2ݖy_tЛqZdoY1H_2]=L{6OzxeL' ,ӊ fJBy}/ҍJIŵK۠6/op*Ɵ%%͌Sh”jU,ڡ Fy'ƍ0Ƴ?#5~BK\f¯voI;|'oG} s<[=sAHEj'zʆэC>˱0cꢃxV\^ene\ |Nޔ_&ٝa! PK~"ܗx<c$1-*:B?GEH#&P4G"닏Qt}*|/u^l2/UlxKgof\֯3l͟JPbGG{,$٩cyGr~[o8XX)a`[rˍ]P~;fU SM]Iyz2ك'VoʽRuio.]Z{l.EFYU*c YS7B0rb$DU VIejo+w q6UÑ;>Eg7̀!8ꎢXe{`~]6^J /Ўg=A\4 (_~yЋArdt9̊LDOo䖗 LR S.ogO糶5pk!^ 3Lz,bp|V/d%o%$F:x}-I;`\~? K#,hdtp݇9zFg*ڢNbi4T%/ Jg0(s eB é C? /lsܥ $wL֍/`]6__ih0@LдGھ,2eW Svb{K$~Bڵf"6CKDyi.[ܬv8hk0dl?.eo]^x`|7ϓ_$y|Vev3H`n&w8qŻft\޵6rcbev`\ 䡑dy ؝̾TEIvҒ,%YVIXb riKÏs5=6|y 2y%yWz?߮7`_o!h*()H @:P83ӲAvo|f%fm h<°~]=[ϷXRʡZ-#Q.v,P?.{am%/% $9n_=|jR{G]z,HzCnvO;Q";zw2y2vx͗>Zj%&3Fٲy$3e"ʲ Єɮi޷ ݸlTfu"(ޒӉBVSӽ'!ƕFn(= $npx(|Psͺ]kn8CeEyl^1xܬgg7nC =v͟v#nng?~)c_j{˄\|żV3ݖ7_MZՓonn]fLbɊ,͟| ͮo7v t}xC*vrwYQoOoSr{wc|u <\]'C~l{{_o㶜mYl@VxxH1^Ƈq# 9LD{RkňYqFq"KfP1d3O:lyZՑn>{Ѽm.{A@θx·rDNL[2#!,֚ e ?0`xl#񳞧ӚA$|IH*w 5Eoa$e74uW~0"{U\u`{K`aIgAy" rU M$̐4ux"81IY(t̃ [ IG=`FL'JmopcB0+!Ȯ0y}-;vTx[z0!籡:ӥ᥻$Oz$R 3 Yיwˇ3yƹ0^r 8D @Y}-W6L˙N c8LiH;HfW:2U`VdY(S\!!G"n vˎ3fpRJR11%#9)$LQZCO}"FЕaN+lDʏ 䲚I1{šAu&?z V. ڦ)̇2@7zC owuT`I"YQ"G4yQؕQa%Z'W){GW 18NtY= r'yG"As@y  zCgIsX=6|iUxKHML8>lՠǶtbdE] zF+}1&o>WzGZfubB e׹p2} Ve9lH|x0ƋP;|'ڊ\gpo;Hd-H[md VI Z:%I.0;IÖvX^,7|8c)F |/^q% {A!Ke k7j엵) EQTG@-bN&EdhdSWHʡzL*5|"JJIl"3!ev,-<6|Z}W(x MU6CIcg7V\Eo438Cl%P"" ߴB$Iۗt[טx5woC!D5ZKzgbq̾V%T!) $p,̳#./g|jIYJmߙ51Q),{=-B KD+Q.XCZQİrg C޴ru:^KGh {ߕZbчtVDgE(R)Iy4P/NqheA4sN@e${Q°^#Ї@)e,}|{f3#(t: B㔺5}@\^S5?/DhKo!=ߍ\r X`~Y>eL8RjavFUN$U% o9ak iRS܏I'Nʃ:"dhi~rIydSǵ"V>M ъ,~{SI:taBۛs4n`i0[EBQI4HB nšЊD `}yZ2M9r7[9Ӊ^Ձ-u 7& 5dYd?:*V5jn@1!ڤ<[aTnE8* $L (Lax- 2K-2%Z;ܰp@J0F9.ٽaʚ*e^&h^~H&EdƑrH(@9 )\Y䜤a׀5ܒtLJvp"˦.r:@$ **[ pSA`xnH q"){%\ 4I2jV;iD^њT$3b)ib$p#3yAk=]ezl;{<* ߡ2;{r9XFB% 4:OXcbSeZL;HVxz _s$ЪLS@TaP)S +pxw1}磌0iyx+,%xƜTUש`Lc/:.t-P?'P aLeK3b8ܱt 4#Fd\ ԠMeĒx5m#B9ԵS]RdsUE%V~ qm+q13*/FI0RQRА\2! s}mvڿLfMvV]~mHIu|psT~1b"$ErƅS&AH.T3(2C`M>IR 6,TV{F,ӪM2a]k f7?~K\bJ>nrv)PGS|"T;Q{xTqL&}C]B-N8q42.W7; } l.p ԕ|Pǟ#-#J˫}5 BX\@% ҷÎ+,東=̈X<*PS5]qZ(q<(0~1I-I*%\4^QWEDU]S?͸WޒH \t@ FqwtWr(,ƪ ?c&k1+u),Ms+1$JM?)$@ JԐL9pzf%ETr<X1dee0K|Hbvrw~g5~]XVl̆9<9ss4R g+3tw 6O4\W (ȕo!Y$ټWKܧ~13AsAy[PiJ1.Afet=OOX}TZP?J'ޭ4iOT;cOr"֡a+yaqVMk^eQ+;9Sw Њˮp2]o~ƘŸڐ&Bk(h4Q@+0`n7[Y$p$'EU:L{0sw;Ux 9G):EcꀬEEdX[(V:r*n-H_t^2ɋL1x(GnکѺjA^qxZC6l^8gaϘ:`%|C-B<.~H1o.v_һ??g~lѪ)n0ǿx;䂟0rs[v"g!P^t"r&[}4%08>#_;]ҮM~܂0IoFxsj_X çb4[z8XfUGPn좁?L^}:R\;t{={ 7Fbr|ŗ0a謁.`8m{V0s 1ߧ@ ($t3THqDglA<两x7qyް9AÕfE~6`d1P203Ff$VSf'[&)A`/%k&Oe-ke)SNnt<gQ~ۻ3M'5,]J^Ec>U^(N}Hhl$U])|ߥ$U+"4(7@<8ĮHԕϞNHU~WsO&%U<+Ua$| %iEuׯY+5L] -͔>Z{MfS#Ez4jf0gRqU#TL E Mԃ,^}_ >5wfL:?لliaG w"202c͂W l#U:9ڹ,u^Pe%ne F瓳ǰ90HV u4׮ƗX cbP9e\ +)=!Db2M;e!Z)]="Dj)y^x!Lp )>RmmXp5 4HH[t/qpIi1XL(F8l&_> CYo̓4D+YWlP5--n `I4NoId.j;6s~7Uf>$x~} в#0 ݇[qO!FBcQGw".ApJt\$@Q{+uԤ;FΗ(Q'Jh(r.k=B"+m%h`0 H2 jekׇ%K%D$ :<$d1@Ěƶm1u0nc8ȆARs .X=|Qe eq+sfŌQ5ض"hJafC fz5iedٹ:08C:ٝBBBhhS磂H@ =:+fpİIFP %1ò̓JP4a,t >0sBk0+lHZ#P10Aj+N$>ȖQY,%@RhV3fn0KH9AL"WymΓ0hsaZ"O c RnP/lc e mDD2At ; )z oŀ&vdPA)eGP-1Bygc/O@gIV.FIbl2Dbq7J91TBÞ k^SL51EQ )₤\C8<͔ВнT;Lͭ8FǏOŭNs:2=Ap=nk`ACTw~Wg54M3a -*e01J+?^t.8J Ҁ)舧Ƨh>6_de@bhƮ&u6j gl bʵ?4ux {t5k3%y8ȨF(ir*8iO6cC '!pE8>B5F Yjk%Z]3q #sz,\0i]2"p1j^["hqK3 !;d<Kd` VҲ ,- JZ5!]y~|Tm, @'۵9Vyp#oF\mڔ'٣VF5J U\ G(!,ygi,lciib@R=YdC8;Y 59V5(SpfwĬ79 6GVCO?U9[Lfwx{J-E75bؕhVqè:N`:|QgY'}]=u~fHpMNb`u`""M[>س 4nӼ==Tõ*Uѽ h`y.m/, CY Y5X} xk,>V|,TbX%1ळ9E%em?Xufɔ9rM}-+>+^̷8Ou0 /ߪ!iB 翱?5s/SBb<ͅ gjGikn obyL(84hLr$_g#+8'<@cb2k\=0/ ]a VQ\,*_CTey{Heט ufao\Z%\R!U(5?)qJN@S֕cn 2/oPYUU@ no2sa/^1Ӿ]YjxbD:{t*$5ehX@aWYɞәHQDʳ;2LNj@}hяk^) b 徵IV.FIѢ9UβMTWLn}n&Bv-…+E{]; ?~.9 _I5b*QE2jTb*H@#%d,s'ůooEłeز!f+oOkݵ\hYzP [l>h4^>u sf]M% 44Cc]HS7-Jq>дv5! }?5ދnh\<܎}-4-|txYpBA`a \0SF~^so4\!E|\[=|&vo 8fT(0IBx !Go]"FlQİz?|dȬ|dc90uF%2%c!yz~|_ (1j<{uTN&ˀYʻܡ V _օ!{0-ޅ*UϚJ+w>(s_J;ü`/_J ;OtĔ~0pm`FU %iHLe#MO6ʚh%ɫwbe|^X&Q\/̖/{[}Jy=w4,Zӥz?7vr_=\#(䆇Q0 3` |+s샨Vm4K։ H~b՝&9#d~2{hՉ(lf@3w3WT`:mI>~S)F72w|P5or"Oi\ΚE5\RK6쮈i<{_~1I43חis+g9sw4{MFͻ͑N$S'0ĪqOҢͿ_FSTStJS〥tZ)ι) ʋck'\Ղx ".=^''8UhɓKI}:[_Kq?䩳*TiiiJb*.g^m3Ň.DR:9i Y2GpJoل 4hS}$*QIbn4ǥS@ 1ABrZ$ v:>ƽަYF1<CN_!wڬuWӐ_ƇS~-ױBS`Uy1Zwd#_c-W\8iHsTFY-1YQ-7z >q9Jq+e\RD9p-ڱ"Y-Y?l_g3a.{g漹Cfyõ#LfltR,yWsM1e q,A+r B1J}=ĆA~qd}%1ı|bUD d9^mAuư 'j/z]}ZvqzT{]Awqh(It #*NSE) b+¹if_I([}OiA Hξ'9Vo?+@8]8҉=),eVh*KJk*^ţ:ur'$?NAbRD)Iʨ\2WhVx8jk !ψ8L<ֿ=Lpv"I Br1%XE1p\ËJK)_0!ql RNPHFaznHޑ 45[PT)̯Y9ظҎ8ۂ>LuT埲8cZp@SOQOmw>"FAf/ n[;8p~Gےǟ`ys7_%i>Bq'_Y3 ;NtUٯ8!ޝ'Csh ԠpvEINCj'A27ߚWFAw-=Ye2>DBoߟy$g0tok t}Oڟj9`NY /_P%i ࠴36^ڽEw8b@Į vƯcTz74¨Mσp@M@s>~I$^x=]pRt| &~|Z"ΟAc:tQy'p5E3 W+Ȅĕr<,}oIf?Þ͏ܸa$p9̩GncēDU,<(qhqbmE-җJA(5}\rl8靈nhjFF]+?n(|Ϗe9/~4GI4,9gտ%S]g[`(e+~deFe0MF/?s9W7 vNN%qbO٬WGIHu}|gl1ȜИ lo!wä<[콦LRM4k!eJMb-viv&CrwHv1ϫ}1-&nnj硕7󜞦oB,ǁJGt9<ǒJqT [>y<4dn=1 e9H;{W֑ FO/{ 4N20JHJ&iIv>(R*nb]^^ʑ)1UW9β(Iy"vrs:oPncQkՔQ/EPO><@xq[,uw>S\hXH#b0~-WoĠsY-̋.\H=SYRjޅUiۅCӞk[t Վy_χTű쬷۽xׄFtay^|Z/~Rɵ4]~7҆i܈v[iqn.EIzh)V'\9ǿ@y /rJ;FT}O`obh q#r:d$c42 5R$I¹yWop !en2_e=Y$yp sg AN+CJTpv`dMK2c42WoW+ -OvLȨa=QN&E 3H F$`NX ٦H$АG#BAX Vv{т@]'nz^O`06O,"BH"&#L'AyffNf7E4t6\$^ǽv5}V {ЂB)Sx~Ee£rg@q= t$iTh:S;ɆЧԲR8 VtS&3~hWZbE:pAi $PD2Kb[&ď9C)ǂ ws"*9o1|>|1dUlPe l\pq4xp&W0{RƗ,,_zn8,] j"^PHdOp`׽w}?<ގ{.G-]s^g{x׻ @ZUyFsD"YsPPJ!(KzSZ`l9sRHqhΙ khm18YztTGTFIJQ0p X.4"TLHڥ[<Ԕg pֵ|弁,cXXT.}KioJGYf.>͚ ѭ 4WUAl#&E=E`g#q('Md հI1'v! *ndcQOrZʄP˂]ʶGoG+rGKz -CO{wonBuOv(*C^?%n*la;k#h7l^r7.&0MկkHHsTge(3B1/auok!HSXyHK?`O =׻3ы{~M|{2/o#AJ}ܧ _ βw5᝕ޢ72PnQSԈAxF IM ;n0j\`^BU8@8{0[i0 Boᦜ8O^N7OB/`]l E,@ oW"::rntG5 iIp؈e\r8(ait'i26͹5{mU8v7)^(etxʡgJ \.Yf`gn\-h{B7ƈRu/A9/m+Ir=釿?[؉~}oUy7R~ia1,oϭ#mstuNk"fjSM/"JO$*pAE籱Yzm9Du r3o !X -˚k2>,g̳yVg p"#Mx(y$"܀@I&Y-NHǫ_zMh ~`eN\ֽscQ^z/'rYpW/@[?'cm& \T)<S8뜱hSڲ|0soW\&-V }kӏR<^ :5AP9e ݒa9RԜG p!LJ/ek WEkAgw}0LAqU%_-i Cq(:}BǺ)uOk!FuTMx4Z$#NG^Sƒ8iJi"~~;\XKKKo ϲ7Y _JTkg{V6iqby´$&T89qZMt+\u]p3(9޴k:Il*zI'gz4/L0ǣ}2m'fp>rQ;BX! Ō*<(I3PTn^ uDžQiFǦL)RM.$g9eXRof s`$R("'i$<lt}2"`gJ U I~8mهvԙ}rvnn}~Ɔ$K9jd,,y?7(^%~BP{6$Q0h=LM+  ܒ_V$) +~9+ }؞:kgF~e/QR/c0E~6&p>0kD uJKnRu;0^){LPÏ?O{YE3 BqBQKJE~l[P;y?[R_J݁Mʽe]iV#[1MZP:ZmuިFۅ~ Dh7b"Tj_U&w㽼m %Ĭa R&in\x$jgcţF\46O0ƀ=,~ќ,=4iW;+ؼ*=‚S%[o^0aP!}iNNn\Gs@Ԝut=Нrۛ}uvK 7g>gab*im %' ~軿P*$?O?ƣ>xھ.󿿏9Ȣ_:vZ[]@-i&h]ozaOMSN? nf遼+^7yyko0#aL{]?mCSN)b^$#,uԝ̯Op𺉀XBtyRqSvh舱kk)`"W ;A~g9,A~e@LS7aTD''0ҙPq0:Uu@J[V)m^yuhz`)B˂["VwDو@vm%#CLTȤeeUpTt"sW.1$ XM_ %Tež)LG;lU $\ϕf#`RY ʷMIvRڧĦx2Q.Vo~+Ow"FRq:bVSو QA c G0,XI-BtuZc/Pc ۆQHq't&|{x GWp70.ZT"b̰Xs[&H"$"EYc@حX޷vrgzp(NIs S 1h]Eىݸ,[zA6u 3?*\߲.),R w*i^!.FD,JkIZO9U.iQ##mERT܃ގ9n^wʪm՝ލɓE@ޑ!xn4E)j8Y.!1Ji:%p5pyM*Ekng[hGEd^2)^QQ1v)@?FK:ʆ_ 13d{gONk+B4f[~Ici`-s̪,li2璲/oiG)F7iZѨܿhU%!'|(}8\C+-{LI/wm~9Vw s@H<,B_}Yߢ,{hY7ER$ȮkUYN~9j3Zؤ,-x^= PnfT#1tFDc;s]i*v׆mq;Nbh^cF[lYWo:qұJh<*THZ`] JrSXQ́jfrCJoBC35#2]eid&17]hr%o"I"%$ N@ շ{>1=ӫߎ;%MEm9$:c^߃ahAwo?(Թeu{zwaCVט?QNaJnaVMfiNڭdrŘ]-;T崄iavr>-LKh 8t-u_.yNtj%{ޢ~vqU]F ]E@U@iWAJ ޴xŇ/h@znf tkZ ҍ>l (gI[i+z^$; -x`u>kO ]:LlVN6IcǏ>9:0 0< HJ2 }i5+)χM"̦˜XOܘ,|}v&G;a<)CQM'Bcyd Fş]'GfuSJ >r@+sj=uzs/:xYL@7,Ok7 Q傮3H(\5 $$=OTMmXÎM/VEM6yxݕY'\3kMXY3M#k B盻,O w(٬iOk#.(qtՈ9t}Zd'#J[:GJyE !=0$cn0$XQ*H<#Vg*~zuu}?/ӫ/7>X?4~c* t>#?.Phlq76 cz{w7rZ_[/>\Øo_KCBJTa|9ؽʠL9V;h/,chh°]:aba|83ys˿0ɁmԍU9}KY5O$ @;+JVFRjʃˑkH4=C'*e^'3JSdvG{M'YeY([VyZfeGq6wj.ٟIcBεTk:t83|oK]Ǐ`9;kǧrН/*0!O0?2\.noN* EoEJy.@ db&$)Q4Y6w>|t2(uhaOj~%ෙ}E/ASfssN#zw2;>8 S3CHN'r+Lzb}pJB`Yp@g8´IꝄ5^9:ze} }Pi)(M *(~96@uhOJ'X76[G#3ZÏn,h93t[$I"JӰNx7(UzVAWJsn1rUy%kНgfd{ 5Z0[jM7p4؀~lzҜWؑsR!7l(sEl'snち;#3D5.'\\jЎ`u"(hg*W݈{٬,ɡ3|C #5ſ2i~}9]2@GBլN^usz lYt`"**\XL`j[j7y6Igg86Pb&~lқ:gC씕~myXE@h. F:*cH^[rmߖ+eVr9h 3DzogKϊw$fU}|hGOk3La2E*-<`MS\FG09T:O(ݕY껖3*.b4^>Ӧ,;@s-43qh`py} ;^ϽOM!EZNLYL+ܫ"'9~ީ/dh ʿSsYi8X {cCm#N.t*z}۷&CJ‡lQ%a|.p%1faGpFfDΦӕU\R$ӪFp}[8}RHͣ}$p?ͽRх (kミjˋL.t`Wp1@_Zg 0tƸT뭧kg`me'0$ʕI{ԁI*8!jG8K`dbT\(ը<} ݗxɕS҇@E悌A=H`,(qEW`@1HU 8/ʆio7$@Pu3Czu3 LӌOo\}>_Mm^&rhx?9yE *Uo]_~`:7?N?c9Op"^^R{Zǧ\UVTe\^b:hgc9ZC*C ZErؘy\ faf`91ZLNqAbŞ~b,bл\f Q͑\~abFHTuүiqkWݞY/E-|dL-;Hͩi!(_q{]/(~`ku0R:4 7k2CS`8eU6tOC/{hFn./z9T5lUp,zx#3yV Lr)¨Ayұ`4)r!05B2h$:˸u$2?OߗP<}_=60@!Sxj':A2Qzۺ)4H@'A+é 8NMQd$B-} wߖwWF6pu3589z3ٯ^ArhCvtݥrdLY렽jviغRzMa$LkeVg +-S%ց u4ϵIAחEF,Š1CIғL2҈36OR)kdX*:<5.WCs_^s!\raھz m_o;Wم7Y0oڅeͱ !3HP#y ȄBldƐG+.z s>Fϧ4yMq>MS.1lRWwq%3bll֡N2Ԩ:=pN:/~ܞxk9N`C>\:l\;:6kF 77u ncCvvH4ej&3峋fR_.OBM&=4z|#wÄFSVoiń$L'7&a&F+z ;Oq'{Tد93`/^Ul6*I&9@Ir>7?ZHqAK%Y3 GG ;JQ*`?+yӌ"xoC?o?Z4M wODGKaaī~̌`OvfVWeT `@$ >OwH@8F)9 4voIwmI:NyPLa0>0ǽDiBH F2oV µ( &4KW/%̠IXnϥMU7UjɑMazHX,tq'E2{bVi"B'#:jRԼ-xCP%PSj9Jܣ/^ח'Ͷs+*V VnVz>}Qˀ#_] KARpEE+Fih&GP#(JIpAD9ETD}S.h^2a?][o\G+F)@dvv$٧kXIveI >}H~E2o6eUd[+*Qj!՘0 4Mp15@Il$vV( 3M 2J?&/j'+XÇOfڤVБO#[Z{4uYL(GybEYԺѴP$9DS $@刖䈼y2diVMo}꾙o@6D\jM kkz@ ieX#|S4(Λ/E/]suλLldmW]\2_Ϸ2v<ǹ7G?ͬԵٚxQȝ:>G:F&ү'м/:<:(*, 9w^?Em/EyTP:$bEOP-^fOz.P.fv._z?ˏg2y9|2O\wZh B0eTӟ^.YHF6Rjm&gh!P{F~sES4/) +H sƜ{L?ýF^S\N#vEl@ ΊHQHpٙ/EΈ\3z<R%r(]xMC"k` n:IoKbڰ™KGJ&DMڋ'S\̥h]*d̚Z,lTE`XE:CQ.=SX{dg1sohB5iAUn0hm{YØ0\=`8j!0( e_EJ4.R4l(-~xvurz$fs z?xtq 9^&hz{;[F/JUPQLQS ޫmj肉dFƌV̶G[0UDa`] 2Z[0`g PtQ"]RHE|M;5fL[3]C:A;t%X i1@ 0qWfKzoHHyzUX.W']6چq YM ߠ6ZұR 4J3Fh yR0H<[^oPy$2tBCU]卶@K_&{c{ږo%.H.h$71"ĆnR.C<8КfRZly% :4P% 6ʄX`}gEo0 Jxʆ X|Ld"Y%)Rh^Xbr!0$; *(jrk֖)h2˿\.ع*|uL(hKb9+j#;G,M{;Ȱzy,9[m9 ) " oV7 DZb?Xx p<ƾu&Mv{,bǑ)a*,UX7zbs$7b@і@ Up,Uqx0pbBMBAs! F|s$GV`+,$:b慓se!xV8 PTۇba;4_ TO}Un"ᡝIS_PX]}n{Z>ƙ:R:5Tԍ-4N@m"[q)rl@0dE+!' :IZ餛cdON^ PAnE5' tMR։(z/7 ;<8.O?o`r)#ÔGUU=6݉M( h+b)0,j"4}:;C QH-flZDyN sl,q @yoW"wvW~r|ruF~`/Ap'UNs8/ 7kz^˰of{8PQbHd(̼2ZcfVHQ'oJVEh1:k#h%ybYH.fhN||>c=DBy;ߙDMF]]^ߟ.l= e`W]31(pQgA5X-^{ EHDj[:O"dl1w{Kå}ԯS@Rqa~V̌yx+/'52^0i cz#NYnc6v x G݁\nȵl¸?邙u\ND?k1Փ=/vKԳ-zO⥺=r􌥺}|v :Nf:๎I<[˻gvEs Wβ BCȾ]4-+RqVh‹pqnܷ[kN,O3/˻ Q/ A(j^iY'UciB `&{, nˆf݈͜Yߌۈ_ڷJ`ԋC*¾ďgK1BA+ z(8 Æ>k20WCn'ڬ^N MVfL}Ew<$L>}0[d{MDKM$~F~6оsc+ O; F#7z$:yzqy5CٙXCyרρGϜ˫&kmSqtmopަsMCw$; 1u ?l~͙$WtprM}TZ6_3ewd0 $AIw1qgIe妘,%ΎF־,($tLMg{%!PjŰǴ,iŸ9 hfݵ'MgW9 ? i8}Q8VݡP ȁG|s^ ?(޹tZoXތy6yy1S.{Xr?k#|"tRVtrQHل?xnH'G:ξv 8gi snpqAȁu6Dm#g9}/ae$rƋG fcCp֯j6 Fޚ@Hr092`GhPb"1yf? MH@oGUWZءxUzGu0&ލӇlݽع^|pߗ˓N Aze,2 `{׋kSփI{3OQ?V.o@ ތy64]LqMR0YlKsn%@sl[.B7WsrvU[ lXme {aoCC` p,}+R:sx38:F G6ِǰ7!q6[cp2>7Z;rȯ2e}30ǒ1{m1 Vl!TƢ.@5V{.ZhXFܨa&ƶ-f2 X]!;ұDcEG߲!XjvC:|^wTvk *}6d$T(27 &@.YHmͅ"\6Z b_Pf ".fJ [)gyuYF*T(qc%301W^Ttk[8&~뤎6^VkLkLzuFNwBpA8| [fv;; "޶` @NV)kUBU*:[v9Yv]7yq>7Z+Sq5h-޹F$cIObv%1qGߏ߈ueo{ZϪ]~99[e1TPq%7$VQ\SUח&: ouBMQn Xc+Nċŏ4r!EȶpCdIÃ֑f[tm6HEݓ r3d(F^y[8:Vd;8`@Cx!w:%5VlYuN^WA6^t7\vȢ {aD f)?N~_n-x[Snk/Zf60䆐 &&P F$`3UH\k6cu7:C<A󘵇2.[eg\C>=Hp ]7c`A!L;6I@!|ʭh|AͭkO2]x`_R`䉵ggټ^1͗_o7@IO I]e碶wܜ9wǑnF<>gFkxTv#t#+P%GpB<USg}y^kAN rSoO'^0bVXXB`U=oC4MS+hj, ԆkRnA 2Z&bXj6 ʝ+t3Q`/l5O|`Vh0jzG }8{7+)y-t! @u//nfXAuq0fJ=؎ީr=J=|:Fh[cw7\sM$oԚatlfK,wqH0a=%H#;V;cf"1ۻ KIyݔo*6T7Ģe u dd& ^,{q(<O0RDlTFVTgө}&MX̳ \RDdmURbB)!]yMHhl69Wqp_}:YK7O߈5̯1UNN+-pMcNTyWRteο!KLQ /:PīK9Qט$$\I K,1+YM:$MHs͏?F3LFX}P}=Jh_Lg6o&vvz?OpLoEy+fs6}Ms$hz .kTߜT='Ze>k$HX\޶84v[kd=oíQsXƄTlXkbH*`Vi`asذ4֥YLfj-2;+RFM5HeVE9=Al+J2R/I\gn! #p"I2I4&A[7#7W4>"vG83OG9(kH&laۭ DwV|=) s}RՊ#Ջ0g@T䉕{pɨu :.s,$m6AIJa]tI+&*E$ M*/BD؛3[›}}H%˥{AW3ǘ?oqZB_lkma74 "FgG w+7FrH*kFpi1i$_U&SߤjKZ)Qyu&rM>xW˛A"A s\DZQH,(jb |erFxu1r:h"rgkc 9Faz%pxU99Y9̒iJ*IO*w/:f-Q~Ĩ-X\Ł54,:AQEG>K@ Ցf>Y4%Ѹa5H*ZV[lZamEoܡX,3Min uI 2+r=~CBWȜ#⸲lbC8ջoy$ȃ+L$@BE=xR˭u5U^} SR('Xv!O/Zmְ07g^d4 Ȁ2s%тI|#) (tF y*[SmĂ!su_ td_B?'Bv#O&deƯpB?]vQBC3h{-_  P2pXR-&7? a,eɭ DLzt\f(NԎ x{*jd^8[hItN=juuZg:FMPʍcG) |g:Tc)DZ=B}d,oiMxjm(^skeD,y o/dџ73Q]SF@3?'[Ozo/ 5+9N*:zRnTKNlFܩYdܬg_?z+=Y[oUZ h cbȦB5]&4>g#^(^*RYO(JR'ĘXRcJ*[ l>؈!ۧHr4JH5V ]Oׄx8T 2,R >W#6N v4,85íMkzG :0i%Ɏ{*ҢBqF"f` &̢K^ve=a ):0TR@tԁPT_!&""FRK9',h\@[<0%> R~i $7D-vhE9]$z=AT#G ᨛ~Xi.&i7Gp6;n\ըoHCXezk"ӎ&$lkW}O'9ا#5!"8V8 zr!vN#N_V]@ gf2cؠSb+yE#2YMbڐ) IjϏ.YX/J :L[hv.{@Y]}XVcfpzyJ71R9M@~eoBgb*!a:ƭL7tl\g̮ߚ؜,LDd%-fYH!{ e/"דxӶ&%^cdJ{bPDbAR#d 6rO.SN*bI5ЫISA+IZ2 Lΐ a+#}u=l|Ya5TFHq+P)۲$i2U&=Skw"0d\M,D2ocNG*V}/ZH 0՗g֛;15n+6-~4^%k[>@$jI3B*LM471d9O j/X"ҒaZ@EΊ D tHGYCmjK>xrtbLP6.0EŶrp\`V!>~vw"2%p/_|p[91AʭRtLIK 8۫<<ϭF_׷'jI:R J1̞~):dL9&b:I!!g*o,}Q*EdFۣZ|OrGBO&4j? 43ґLu_؋Pybwb~鈴 xO沣6|n]ϣPƓ]{6v7_A?'PY1K <$,fw̆j m2[xj|]toɼЇۜk{3Y~qzVCo+ld`ڹJ?Ȇ1֩aе`P:!T|PXRՉaug4LTBre>8n[/7$y;A!NhdA37}N߷ӧ ijN:ۺ]ZI= [O K@TcTWxV>As]MKJj:OR3Z+TTtγ!n1ρޭP+av>dzi#܀G?=@>nA(=. UXmT㚚$)s՜PhM( /uA7ԏnZ-)Tay{_q@R2l;{T1+*\Tslqӡ >.>qdrR=%dTcyZs\= {TUW},Ȏ]K)~Em3RƝ A;-P:1@OQ3=ݞ ktj5tƚkvS ;},ЎÚڭ9ARu) QZ>,-ЮS}6c! }{HD{%Aܪ֒ڐq Y|ꤎxJNRRˉCG@G)n@RwϘQ[^7z F{i8D5Mg{W$xqZ[zŭqԯ1z9)w@fWVD0-DOZ]TsƷ ]L2#$6ϨRgan6r *NIu2NZѠ+c1;: ޏ-֛͛~Q:2ЊJpqcjq5W[r͟YmG/EpN5ڦ_rGO^wf'W{HuYxŻwW:ᘎMݻqiᬁóڢtSzI};~'joIڽ~mAJ Zg~KZzh3SjwxRC뵏\r^ /!Yj=dFc͇brî6(WO$#`C&߄ĜuJ1 RͮK61[h6G!~'Fj uVeen%YA S=e?$NRH*x/<zC7<`4Ў72W~!ك++ZSٽxYfR;](eӳi I)ޕu+b1X`y`Y&Wʼn(`[ۭ|RۑGֲ[=e!+W%^bG6}hr }B.}ë,Јs^RզH6l%z0-@LRGŒ 4TyjϞ,KX#Yf<|_j✛$8?#D%Ñä́q0O-" w BLQ&8 ;R)4ov-7.L{KD'D8 LBo]!۔"^ 5aO {8W+c*r\9AFd#IV D!EDaS]Gh}{W'Tjʢn XZ"9~%B|'[>H)Jo^rJL$"KcqT65n&%gcp1b`D(éWqy}i ;H!i͵? ;{L\Q.%ñB$!r< _g2NG)jg@dvvg JR|=_7O9Z|k3^>,R{a7Z{I(&)&@K~Ld=X;/vl7~ХHԠeyMx 6TKJ)c*:>w ]Gq ӥfOmLBg_m՜':NʛvSX%rINP_DٍSkk!> 9W/H*nX 8XwQnLǶ6R!& Y)9?܄bONxc迍RQ69X]^]93ВUƿS ƿ{WV y$xQ5FjS:_JH  ̇ `T6x5y EVOp/xz1헒wMe0LJ~v8ic;7t\Jcv7ւY9MrIKt ?7Ef;cEZ/Â-S[GFT=s E}k,١rF׸|!|SseP >CB}\@ʵzO 8~m2P6јՕG `/lg\0Q}~HB3`KD0ylmX˚5p0 4ZU46v"s\:BtPH8>|7yf~qakZH( 7p LաOɕSܪ}_"k LCl|hQ3y17)RU<3I O~ Ѥd\~#fF=T6QdIK&_UEe6+gFH!-Ϧ$jɪl]!fʽ "#ٴJ6b L9jyOz Ru}v7|7QT뢠){O‘G`l D.z)P B%EnBwLs-{)_Н:^WE}u'CbTaWW7 Q3;Wj|q5@_-v'σj+?ZC3ӸfFcǟqs:K7uޭuO$yz{He-ō癗oNՔtr+egN.##bEؒ{_9ړ,E7_5wcA?ļX|];NůF% -ȑ$i&@ &t")7;8IVG-!2dEc0lU fľf,5iX#U@\Gq+aѤ_!nD55@H[սkfn 8;_[,dZ tFo{h>[TP΅w-7#F Kn3\6/6;f53l5o)(onl#It$&>d?e>[ZB֮vA%hL ` e)gf=l~u Iݯc8=~{Q]#Rވ9̀0\. 6q~U@>/앍c]@i9;r(Zr-P{/KG[@~]a_j,!=X^אh der`Fk}잸h2{"^1czUBn'p+^,/Noy|6nov{iZ\zl4/Os0os9lϮF&;{i.0wq&aQm Q>/ W&ɞ^;;~psAtHE Oۅ;1_=F[H$3.@Zy6)AB呓HNiUK̒?O&?y(5anكܕq=NdKP#v~kd^aDc2@+z 5xsz؛aI5 Fs:pPB AьogzePwR`qdV# y7S3gmP&ȩ<0.!ykXK$CuZa8 $81b;J\J俧zvIwgݝ]qILWUW}u*u$XРOYkpE&?h'RǪ'^ulNW]¡~=C $.Ǫ{Tio|)t7@%ikNkGSnpK DvR-:BÖlD+O p> Y[P]/5=9iXklC3NNYt`*KidI9,&&uZo!u:^N۬g+,ZV .BjO%i*}砗Zؗk4Aĝ–ڞ-%]9iBp%x ׷0Uj#Žp}~sZyy3zhaCrA+{ߌ9Yq+fmOMTIqo~ 쟥•ٟa z$N;?iɶ~5=j.V> :hLfo3HP[JX5fU[冒FY] H!gcڱІ^rDysͳ o$% a4I\Rr$p4@1{Բ ltuUă)JZ>Ѝ$D[td=s t5V߸DthY u$佢u}-jV=B `/έ-Z7-6M(mm͗bKs'eŮƴ\JiV8 7f=cSL@- u(j+lcp76Z}zENyIЎ@У@*ڍ@tl|ojǻ?KcX )y&2<xH胶6K;._>:>VZfeM;V|k8r#  oT8һ0UDMi3GWWo@J^ѵhm[|Xˏwy6mZXU_,UV`zmPrIj- nn%9AH5-Z4ݳgvwHAnLIJY٬igg/mRވ-S S&|bV]4u@:};kvxy|x-Yh~(;R;/d %lTHY*ه9[]͏mW7L+Viܖ( }P[wir[UџU̖ӷmKolizO{--/KsN7G'E/>$SSޣ}Ƕ[ "H6щo*K*[[ѭ敿8n5䛿ߖ:gGo/b@,4VFiȌ|-|Urġ{Cc-*vWZTs?Ԥiġ#oN;&s:c6*/ 9^qDt햫>j cv{MQs}GYeV#]~%avO1sCL.=:qFhw#kԡNډN4vMvbD(JTQ!e (P(' /EPRfq-ʩkA=_-@ZZ;YQ;<+Ma@#J=WQRQeD!s:kQGv[)ʊ vr;[Au@[J.KQJ `4KơRYm FKl.2*P>(61Dt)kBD9h-Cr!0ɽc{Ti|h'tZmƨ!Q;$AH-X(;YR;m%C`|Q%; E4!)3Rxn\@mQ4JP[aS-7R8Y v%=`łS(-*˚QCDSbtdҳyㄴU"5*ZV6,;&Dg.`+0VlpY%ϡy68*!X[IB& Y9HGRl"{Y$dCDAEp^g46xjw嵡hSǩ}2 dUrgV$V{O+IeHV:J6r0_r)2tZaR`@As8z-\/gA<[|:0yňSV+/" ^,25Sfcb%Z:I"E]VHYj0;?HncO4>OZB26&]z1#&,=S 0f'K- 64*$JYkG ސXFn*m'Sdπ0rק95hذ?كKljTBe.`5hS'bDY\EȅyBfxCi?HHt6R?4V35hS&_>̻@ bʼnب 9iVTXWKý:ڡR+Mw8# 4h5vlIxݲ ]wނMȞ\dN%nfuIg Bױe)L4^ aU,lEVCcL`Adr?7Yv%h1*ώ16]%mGE5 +%Q&vY3RVs2$ '}"g7deKH٨%Y4! 5Kmz7+3:-dB`1boF`>yBS*]J#sRo9.::z/?櫋:}098ēN^ Y̝I`;hFDbYJr^/ӦgҥBd`zZ 9Wn:.6Ujgg/uvR?i&W_Lp4qp:>ikA_&,?C=~Ps$_4>ٳYw/_k +qK ;U![ ŧ2_M۾_k_~jX}91GC`9w-6mqJd D'8hʼD<+nml/[I> Um.2}~7%OU޽:m^*0oYtc޽{u]5vx_ n~}>azrUݫKy'j~Y;gcsy33lLf'FI ƚ,VgER0L5-`zkbMX?7:4}>* W,A٧dڍEc=*GZ=оۏ1. KYϫ,{z73LaI\Kʸ /wϭkiY/7 0.-xl=x.b5PI޾D< d^& c} _fht'?I"kR/JjǢW԰rqX6B u%vZ<)ckO^u8'd HܔZ.]bRDE:ɵsV{Jl!xTnwVbђg:;e ۯU8w?etonYӮ{dug?{۸lO{~?0`ƛ 2| f?lȖ ə,߷-J,JdEv>U]]umҕ|>Wۯf(l]7},+ͮnthj K0$@Eֱ7Mfc{63BЅV7Eq׍7 ؖ^ k\=eh]?ϦF0oxIC.jl2]Ò4RqP?G $\Ҍ9n2#r8!,ٌvZTu4xW+t2JO?aE/DH(>׸qept{0FQ9jvuv zR1h FBgŨXtWOJiœR'xzb#a{_ȔޚC!-ڠĎcOw3?)q|swq|+پ|+ߜ~]}wgûmy}CO~~5;4S&͢*?>>D5iiEQ\k7]pk-xJv2^3tϫ ]b/)(_z^׃ۄ$qQ:ޓ<(FK/w =cw umQxw]xLt+Qθ5Tc 7c91(nZ^/pjT ~)M2ǹ ~2b6V:RCLdis,hSEI502s9Ӗ`5RkjF2tC]qHDvpMw ZRxf#k NZɄ7ueAP՞R`WHDaR_wBBZ 3v(ᤠb9XSB>m[4|4FVI‹Jov8lic;hmc8;(=Oz6p'8$(ESɇ}h `:CKI"R)J1dXcΣ7[a`M6o qHqĨvG7ySJiΣegY5?Yi#Ο:-  t>sǏ@rBsNAS I>t괅Ѣ&X`i"RL'8FbzGhl:o9pŘ_"Sa8sW9|{Ϝ}(w2j>R4 &/w=v@~)B~ύo]v=_yDtt[<]sfM8>[?>Cmrw{}jko?{Oִ,(s~`?؟[?^w`Yd U)Cʐ0ar gT@#bh-Shy~Sxy>x0b*nV~_^Ïvщ^#QhB<< ~`Brs)>P橶y 8(r;ũp{%#_w9j>%)Kas``݇"ogx ؿ㠆~LZ!]X9=%<&R|,% 2Q+( yq$#aE;i%Q0腱/ -Nw;۴U_j*]I2yg@#Lw#$c` ѣv ;?X t_J1]S SEH:,?28ҕy4t4o dy)nEsƂ}W+)n0Y~'e$^XK0R\4;fʉ Ɲ(/ TD/f)2J<78\ 7w&* ;͠'۳\1*.Sai]f##g& 2BČ: y$%s(;`SDa6 J(yv܍sgb8+`cGu!`Ac"0#3$y F3sgu5+u7\" Lĥ*zRQG+.0_P}iFkq; m|@ k o?{Ge.jrneYGbs+ܧ ;d#Ps`Y>DEA{ ȋ>JtҧpQ@N4BHS:N) Pa|~;kb =Dh\ 6*ѠT?UbCYLaλ코Ofj%t'-*y+Nݶ&?#cwi@0}s DHO8Q"Pap,Ht/ +3' Fk"\Z'Y:DHS}JUi* v튂VEAwRL(II1y摌ER U f4VL1T2yTߊhzI] ی*2frDi],nb11)1^HO9#<$~$qÃT4&豢;A *Tlu`bAk4jDYbV[ 8 01s3)rd].4c!3$ЀQHoUIl%2b!3ԌrX&]JSs0"N’Z4;Bjè!E xkHsnݶVHVLk9)u{h"sY&BLbAO I>H)š.GEX<<ٌULAAAX6C-f6*k%Jf9 Q2cB H+Mܘ\Wo5&{>x'$5΍08`oMTnfS[I6+ %{F Ps%4r@2 Pl% Dfpp [ڔh)whݵYp+JJ"]ιBQI0g0\)3vU[YaZ B.ob߷,z>룟t:x}ѕЋkUN9-T hhSt31j^St3ꖹ8]UN"eT,y{ q4{?೏:+÷zӯQ.p}!vPJq \(ee~MbL<-Rxf^Z`~w~VA)9hT;.;Wy<| vmYWn0\p!~ghܦq] \k\("DM2hX*JGj{$kGG\`ڑuGE&0K5 . P͐Qz`ب-%ݫVBG b>5q۪l< iƍ{P\F̚3;iqO NNzK+F4Jul.(7A5}冦jP@c 7<&Y?XLHU-T}":lP(N\7B1$JYSvzz[:hkۯ~2_=ߕ{L3PnSnL$=޽ !s-qCd* 2Fl;u&ǚvX/) z8v 6֩0 f%&7Kʄ o6j4]y&ʼn434~qozûDBލ;F>sH!9HpAON]X /{6_1a(a m= tO{E@Tk{m'})V-IEcI5g8f3Kz!-׈X>?J9 ͞Ja|ޥfC $i6"|?m{t)8߇% v_ʦbUʗU,lӍ;lݸhVoceJ+*/EȮJ>퇠`1(&nMm6|]Ǡ( E&P BQDx)bdz`Lla vE6(< g<^|i ρ.nsaƝ~/U'o69-\GРOϐCqNΧEAg% Xy^%qtqT'sWh}a=kBGr Nk$h bj/%{Aw[ZXz0]đ7n=M4]b!1 Q(@G $f$1`.Zen#Q/>û]D펵BjsQl)ҷT!Кo|⬔9,ݾժS68ҍ/g`/L7G`ݿ<}F|ܻw? ?>l&?P^A^ /iH­_|}<z<眻boOQl4OJ2L0h^#h >{_^N" ׯ^o D/ |r!~Sf ^ N .8w,|~%@v9f<$s,őgS_A~JjEdGFT~LV"`yЯyC~M8497vL\XOwzKJ]+Po96`&wg_# `n/p?0L눿w]q85KUZ+*yJ! ANy=YQaS-N5ureXe#JIaybVYicy4-=iիEi#歃 4h/fI:tNJY. Tn2̳K;̀Ys*pݿ: wA23oӳ-PxZ|PA%ܷ>`*wr{,\a SYOfiqB%J ay+(` `9Ob )l/ƥ|@zHvnzz1zozG:ޕ(‰Q` d)&&S gB4n直]l,5fֻ*J`|Ag ̒ YApNʩ)E.=Fl?_Qg0 j۝]:S5?ה/P6oZ̥_Zi'PtQ+jQw ; fF%FgIcEAJIGA.0r1i#HORiDYq0\c颯\^]2~(n) yAvFDo,ɋUj|{J;CNhHc}ƉHöFO?[379;nCV*)Kd1"(&TIL:\RNK bx)@)eR*' 8Gi^3$󸐯8Q[ uG 3\z;~F1ܮ̯P(]D",z}Hh2f.eͨZCSLi#ւV-X-,bdžjB)!+X"3Uo ATȌiU 9i 3S3!HoEcu2e9x73~Lb ا[Bla0CiqdA`xVZ"n*J QTQOt ƅ`(YPk=# W+X;:7^wϲ,ذg2̲܊}^;{@{Qk8ashrqJIb|"5q*=toJxg mJ>I3Oۃ",ޒӰ4Ny7;igiTI[Z)`N/>utIPXb1yh֠]n1< 0ټIJ0Bu@DƓi='=H>pwf203,"?F\[HKWTUHpɴŭgj K /C%(Βa v=Jư*AI&i]u*&}܄2oFnH?#i$+h7H N4[\r7𔟊HdVSDE/p%RdRR,'{k3yҧ+Cꁤi*%3j b@ƳߋЮFn}:4/d88` &v2uMj-0{{^R*Qk`(E-V*>SXCgG1q#UNnL5rxk h$Zh,-(LY\*6M3r%}#{l )9iJ^\9\ ʛTl?)5, ? C&Ԝe)*=A1*Y;O(y'"V"Qu @ XMT4O4Orx;R|R3F)K3H-*vȤiᘆ)K)^C~|kVq}jb~tFH &`g+'k19Uݕb&MZ-c@ޚ:UT)Gg\,>yjrqr)Mvaތ}N7(ܯ]>,"jׂJ1Ew~^Q@z1 Lj15 ƬJ %X2;%>nu[1i/A0Ffq5jI%{lG6q"D]%z$j2bJ[ZRE 2L lfZy䥁*T,zCG*NhցVFeο=ҦVZS+JSIS+J[[+m]!5Ҫ4ťnnQT&*@Z7B,rRumD%{K5f{4$+Jr5״T%XJQy˕*&2X)})RM%AFy7bg:<3$0jw7)d'n U= }W@SKRqSX iŒAJ(_^{ǝ,|Rr WijL˃QTPqf+iXO }V ]8#qa ذ.e'p9m:XD*Z0q~VeMxJw$4e\i⮓uQY"0/%7A9i 4T@&~ҡtx6(D25bMI'hN%¨_ c;AKUmJ?UX^ۡVynv6Kd5RLy~Q*Y~r*(ֆI~ɼ"x*Q"F)Jʏ.nZK y F p9Y@7(J%ڑ}eE{; N DZIfׯwwj} %r9Տyʫ_JǼ1暬gšKo*P՘R?tV-_|Whxm=!-ޑmlېžlcd#hrwXRw7#d-&.TFqxoRNr5uMJm^W7AA{=~+Z"*ɩ!p 52o}iߪHF7ҾC͞:җqg[ϓ;I/gdfDq<\]}c.Oe<5e3 ]Y6qs#]+1;;0;o[_*<5-*y:yJ*rPԄ1A/`O0/spͷ͵~9OoyP^.K7{,Mqpoew㋇,_IrVbۇl)$5PRl^U *BFfUn8(pu TnݪӾkg4=T~w2?5Z֟:[COvepkn⚻n>8qU\dL#.O*|,}^fjK b)RCBQg ,ww{*^rηljv$*Sc.ĝ~0_o.6zp1UW p^%!Ek(Lo#g6^A,4=4|1hf=m[4B#8/4PŞn*Z)_OV%}|]wU}TEjb "@!&Rf!ֳZlU`6*~(T%j+̭DMĩ&GS5f*>^GhAS!#+tF #bM&"FHesƱ6"ܸWv$@3n@˖* dqo/VZϮ8SX(Ld 1f :b :C()X0ưDus!{՚ֳk39EZ>VtT!EjtUI(_5vz;I%ͯơ[%cMX)&b1|eIJSr.4nNWz!?jIGEQQkP%L*lme9!RmgV)` Z օKM\cOPa4'k$vR/ʓ%CSpQ3JVJۼSpUjO. 8H126ɧgnj>`>m#ɧ4;l T=Cl4+JCVfZ?ݩBL)[y]xcд]Eќߝn0#K-^¿wEB;9bW1`T {9_.D@t^c)K- BJ['Ж eGJ@ٶ(_;e+gmHGo9=wekg_.mOg/Yn2w夘ON(E>K3%gӸ2vb6~:O^4{3'a.W%gdW^3 (\dJf'J57k2[aO]g\-fi\ޞr)X,^<(-)h*xDŽq1 F,!g^1FЈ paGB{ʼ$75g|J:}xd{[\8j3v7v6p67~QHYQrY(r319 :8)t5סR"8oVE(UHPXR I:bH,U[ }9CWԺ!z†ᙎq8j\H8c{Zںm;~鐃 KU(}$g&e$Vh59s.xiIγf KU)G/e~©$HDKU8 da9KVSSUt9DmcpEC)<1YԖ7UIeRh36ʠ B&'" * ~le1 A)8Ý1[Aϴxx#ǐbd#G-aexV1_0I"LxM1 x Ϸ#\jA1Cg4^1CfGoq [l!Ʌ5e;Ly <ƈXP҄Cc̓1N #q:"WhAtV+#J9v}S2 1=%lpMUK*WQ`G? (//>rY5&b'ӒC{ߧgg뵢Wl&'a9ݴǜH0WXЮyn #,*,1^C R5"9̟|`i1>@%rID6L[/RK#Ac"‡-{OESAbn "Q9 v n{\d'05upêrq}Jcmw;^wncܭQ~(vVwSdajQ1?[j=FK4?t4PjP%Y)JqIQկu@\YC%ꐶFR,SB1tL;>fkAǴЎvB;žv$#mvR.Xe"EJU0AI\$u\Qs?wg Wz|HV+G[hยI  ykT@ <ƀ8B [H ca3 +j^'VtvK:2КN F1p׋_F1ҍnYu6k]&Z }ϴXȄB)uB7? >}o +;L6`H[1Y(~ ~glYׅ֓*rPT#ApI %P[o^6$CRlp:ss2GQ]JǷ7-(|]Ik|鐅J)MHju{,]f8Βt;-JEOJEb,hE;ca>e+?agd{+8bД<h{Y71L:b[}٥-'BǏdvk!![xx4;@ޭYw4T W'Aa iᴕ&p$QaC9p2]+'W|'Z6G]norlq7ss #|_i=WF:|C_j-^lGr>ھyЫ~\, "w#{+Ijw)Sa qyɗ@[j ~Yw6&jjhGt{涍, _2YP}զjmM*u*N&_bhbL I9^R $$w~zo=gc ұYUO0-46>!a igdVENIm\j7Βyscx Ϥա;P TZ>/(VP5cmcFGHO1^ |<?߬3 ^e' ~,x N4"9$"D08jb)_ӹ[Վ6]:Xb;e.u{[(BtWE`j,b9|hwŲO=.&q.b5@v` 4[O}_)Aƞ0lb} 7= G/VR4-n(X$`ܕ]0]anĹ2 'YȮ$R}Es|kB6Qi6-q3# Ps,*252˄&^K܃o^nt0Zg?L ҜL -L]}>(p1 @XSbFVd6Vk V [ WVG:Y>So|w(o8Ju ƆU":262626QFfaaacwGƆextK ./cVUPHذ66,A+0w;LbDpv&KLEO)-,2*MYPW`E XL4*H24$i T JΈ\ GFXva Luz*✂S*8JeX2)I8J4F9.=mdR |ƁQg\ .$=I[;Bʉq$8|ύr)uxgoAn4{2 )(!9ruOJ ' ao|!ӄ⚜E=J"ER}qh)Uq9 Fo6"Op WiW FL&*eA}.v@n;:/ %Դ+XXX\dPq"ĨL, $Rpp5H2\Ŝ4s_]ȫ~w.E9N6e# )@hj81O`qLg#Ce QId҈b#Xǔؗ=qtSy*)HY#,k ?q9)#ɀZ g &PqLBj+9MCepP$\;ʴqɩo-FO߿l.z.LR>o RJg™ (pt,vAspDݟPT*4wU-bG(<}D2 MxEp^ca>|ihq8b\&,dFPSPө2)*A5V!קY,w0ڴP.8-J\ht$W7|eqIaLR~i%ň._v7uY`x8)Poq_̟`|(E#Cb&82,5B8뙘$hkEF,#@48} .O9zM$J瘊/1͡G!E =UmELvu v-zDx/[.%.qxGg]:<>ߦU  ĩR(e4B*1EETlj=(%.w{Z}Br_{vxЗU0 &I1c> IvsMo|,>}f~A I>]DDoOzo=g:6? ?3ǓpZ|vQ}lt/vnhs{ʨv\juT>gvo:Z;P>ry'OdBL˹jYTj&IE "C&($ɲLsK H ;#;ZZH,b|'K*B(؈4H+^gmmCoh3s 2!38x.-#-5 rY8{_+d(o |P&: MPG$ I+Stui!.]~rY R&+n pe&Eh1SfϜ*Ix; 7yTIq/'Q/ ٬zJjy ٷUEJ2[Xcݮv%VL7L7L7L7(3]9' ǘԐJDq+QSlFNS܆uG9.滸;oIjKUB1xt0'} b \|ɹNxdGIE:AԘH# FL$ %Bt/ )Y`Wx`Q+X7Y1C5hAe.=Z&.Z8XPS vRQ0BV(^ywrcܻҎ,A.5]a;qԲ˷k1lĜvy%@m^`[C0`&A^(sE/swRF|$웟(|yN􉃴~ék(d;,|0 ނMXfX㠼_4#owVρQ;"8+?őjŏKuܪ Xf&EKyP9ͻR]PLы=쪁 VanVߥ3)DaI5%RSRD_jJwoԔ\jJv:iՔ4ԺԔT)5ԔT18pЗϸĿ]%ZkI 1Y?)8FWKnxa"v4CvԶ1pHUTp$J!kOi4{ic(jtī5_OӤJivYyWߍǻk?G gPۡW4z'%q{Ԫ;ܩ^RRKuZ*nNźg!3VM?XzA!ҥ.lUl6z6lf~T=غ9S]k?&5F0VIҭ{GUܞs{4{!&-ߤE}ed=,WBN+ȳȽ3FQj6M"T t"uCŀZ5cqQD'9;txح򴱫FQrz7w[l k'X)cqltK8S B0(A&OIB/(ɔH"ω4j%Rc W14@0 ^>U {kArS[ŐZ,]kMoUBbӇkT}oz׻߽,t-?]عھLy3VXX/Nؿ0Nðm>//v+{x{EEyzG"z8 gQ/*UnV]cM%Z*Dб3[]ur*PT ƙK.+8p1kTWΨ\:8 gѓy*u]hܬ8X%Utބ@C hA׍O:l^̓ Mbtp;ZM⇟_@Es[Jo{}(0$"e ,*ه2(g!S}db!;tX4_Di'^JHNt##0ժcJmc4uqǖ5-Ihv.j?wp[`>[VEdo^=%?(GhLoDIL0$F^*!!$5t.l |1LW1ɻ?&đX6Z*HG(WHV1i[fU{c!*8hHQ,xC0,Kz48(288m_hw? ).;@I0j$oٳDHC55[GkV,jf,j~"jc55kk-a*ۓJ+p+hvz['o|$@s4K_٬%VW˵}'txǀwy1#)Jr;XKTjUR3=[5h^'M%#xT3jH>҂,&H,Sf25qqn UpRplS r rЎLHOl*<عii:X-B;'m;n˾uc.F" QHCyL3l>ܖ|V)da<1n|pJ17X.dQ2[+77|SLC}JhQ7zQȃ$a=koƖEה``[7)&0fȡFt%9{$e(+ "y3gJmXQe^ِk-EQ(; a w1c"N!`/98TTKj{p*^lqڣ38)?Z)=Ntl|b~6%s֝6RY$SIA)ɍ?fmZD_OO:q2Ls0 \spA_|o&_peX|y;@{R .$;7($B",O:wB( AVitBDG<*w+]YX Mݹ+VLk&uBpa#n{q rR`Ukz9Rjȶ>ۺyxy'54&ƅe3;@Z}~z8g|JP&U};Z+vD\~%a}pǻo+8ưXƢ(0ڡ~mm4 Ke>q1M"PJup=O{!}]R+2rx*Ż.\[V@`XRk+Mьu|V%$H2[w!R# BV-&, *ID:%h8)lZ+ľ.kdse0QXXX距_%A<_gy OAN*EiJbwOB7 jn.g9s"dp-C.ֈ c5D(LciM4qHqlO?W)}zKs+u;u?gH0ڟ^IT+AJ:by/'ɕѐecXPiBmBmv^7TL80.d icCE \ JU1|;q].cڴ*#Do43tG$\Lp.4V"nCZQ`\ܵDW!T}Mi4aBo$ӓ <3_p%Cv`1īe}7)9Ƙ N8jTdii.e}%$u'w wO4j5\7~ r*Pk@^y)ǑbP|slcU‵ߛq,/Nnj4fws[-SngSC+(*LpPS q갩kD"_%1iijlQ3N,OES@-،pVoA;Miѽݜ{yU'- v.fs BG5aJϺi9bWVrX(eVkb.=]z,Xq =lU$!RQB&a愱UDT$bH#+cwL5ԺZ;(*.ĔQ7eVbbom%܂Ϻ\1I3Us$)BKҟyCgޏ>?NnU+jɴO&^'q&eo ی:L̽4β?sQ߄.vkrm ?w)P0Gj吓~HB[aYu UА9|Aw^O'20u:Cz% r0~b sQ/yekq 6)7vS|_gގ&MOs^ ]f /]f7'ٯ? ϓdx~dӾInGgiҥޝoP\nS_Šĝm|9ނ nmz+:U^Ipb$y&Kex3v8h<)v (_D)ONjo/_K+w% ?$K@:3#lxxmCn>G '@+7>?5#csFP$a~{ oa`Z\y2bd<r^)5f~>3?u1 M?w:ގ怿0%Mz'(ݤ :ތ]}2= {_N_M"9 _ ˮ__@dJ?  )ք^&/ι8A&9(r$4베3ۥ#ˣ/ȫu6(8VJ].³ea<"2`@1 IqySϙ `;35\R? ,@ߝ׮xmۗ$>w[dXTR:@8 #Hꘗιw!¬a5m mr %Bkٍy tLk1 H@Sp0+E##bjBJB’)"u֢$5%UXJ?$ ԑ;GP2aA<6gv: ,Q8N!XԈ҂#)#AcPZJF]Q)&^n>٬):%cG|hF7*JbLj I68G%v!j3G6ӱ :ZX)BD3\k+u  ywb%&r+k^ g*KUU2k|BRr.-B3Gӈ eLY3𿯮׎d6.;j1lcɑE,y`  ݚXk"j.@5aS-ܤ"#+! #I9ekAQyR\s$ZnPEZH-%h 4KbQ[Ŏ!ttҶPZn|k[PMas҆GVေ( vqC)([S۾E3ưo ur,xK ZXƅL!mlSһ+(*X!ɷ=ha2 <44-,{5_<yzh )K:>+#@r52X5__c]F1age^-`\-T7vGᇏAh{P<64:}Z8\IB"{+lIi`]5ZP䘦Ӓ{)]$yj!!ARV ZoL=vZo\#7׶E'^ntz^$r{GAWvEk b]oHW9r,w 9|4mȖF3Y~դ,Q&(2ؒ(~UUrmYO['%8dF(`q&i8TNOXDԀg+1c!7*`lQU ,:”HJBhBtN9T}vHEX) a2o2.F0IR,G2)qb*L^5دw& 'WM*mV%f:a^7djF00keqSiSLRa캒r)KR430# 72D9a(Y(9@2+@|kQ ɉYLVԋ)'&GcMzgVlj :9C&Q(b$N#CXXFa4i(wc.V-2TO! #e xpfXWPPH"B(3De n$T1aQ,[q O6-%POJ+ܕ$~}zv0Tqfe3")T Bsq(퓳bIl뵡O_%F+K^U,Iu}OuVJY-@Qn˘ߡЉBL;64R ZkMeZ| LwӱYT57v{k Fן?MC8M"|ˮtA…x }߻\n&| \ y%3/è%!fk஘C=zM& w{< ICH7|:Q9 |t'@ZUT!5XBCCZݨ!Ӧ|4NTq,§:clI݄0Q0}r" b*9gud]֮cտe]c<0ڷ7[SV[ۺ=3L-/T^=e$5~$Q'Rvs޹o>j[w18xi韟lfYujZ"/g>G@B!m aDspBY|kʑB* *(P;l&nWsxַv[P0lp&ORi L&qO= AHDsZP?4M />5&vgԚ>8J8uvgL 59qp? Vq|"X*tUC 8w{#)n sP7T9 `-j ZCvFcԟsDWÌlsT?f9ļ w֬:Ǔ}e Yd,>N X yxI{#>-<㫷s Pcϥ?[B:@! (T u4d*X }sd%-uloIH 3I2A[9<Ds=}2|]U,K&`O0i@9SLi>L+T,LiϵYi77c,`2>f[[^8HFNpv9EG_,Jb.v"^KjD|;#W$1źC%<2"cTW*!X|Rv`*5[nWhYDY zzxj[ޚpք/pij$`0!5WZiCPPcrɌhB:JhEI#Hi0B%#S6m6AWev(Cl+wЮ̢[<̷QM $;r0pH& (%ؤYЄTQf&qXM&21)}]54Q[]*JUj a G*B8eGN3y0UPQ$\D:*o5z6K{x:3,q_j_9_f2}Y\_,Ox\,K`͖C4J1F/M8^f2^ tB.hĿ[=^|'Ӆ3".O*bl3`}vMB 9kBATi)1O`i%D2$e)a !'c-CqV) c׌8F#F)T%!10 0xc% l/[^,sF~y]~ٝeey yJƔ}zk2J 쐫ߒ_pކhJހe?q"|^_DDHY^Q6K#Ǜ Gg/x˾ÇK"ob+@J /pn.v9%jv}eQXJ d΢(I:DP.ȃs۲{T+5Xbͥl# pT>*" 7@*WNNup(QµdaWCRz( Zx8Rj'#R N()XFHXrO@" &Nx(טbqa RJAHR#+S-bnZݻe  ^=R0{y3Ԟ2^<!ĮfɅf "`(܍ Vl-Ř˄U:6${9 _D ˉ}?\% >N!*&0 NJZLBV`)Wvn{rAF,&Nx> !':Ds ޟf (F$=GIKٶ‡Xip(jZߗQX_:rQqbHt%wo}ؠRdwEV1?b|j֞͐U+P愃=c1V6XGX- N+]qװkIybL8sOu.ǒJHhł0&H#LT1 XŔ[fSn%,?x\) 1q>h05 OƂVĂZ,§Rc[RWP <&s0x) T cISн(BI2.@!2E, lk@( [/ e|(`ϡ5(d `(*@4 BCT(ގ<iVm4x/s _(&_wo7m|;*z{L F! rd`wF6cOjv (3ʁ:d@&3j]H(_]^:oTψ&[,W窂sbDkIMGUyڛ||XdcR @lqb=b$ʕ|鮒I> _.iΙ$Qໟzs~}Ky"6%>V\v;>ny+.ڏQ@-pk R&uȴu'dzI\HpdދuJKbKK~|~W9j r=MwV&ZU9:TK!P|hX'ށϼ8g]sيZji8yG/=_upW:+Qږ\9v;#8l.>-M{]6CkG.z=U_\%Nq p[cj 墐ْi:ux6%ݓY(,DoOR=4kTjۊr)Dbm sa(u`6ֿ,ٶ6)ƞx.YKqfIWV@3: U} &w{dָokMޡz4M9b/>l_b\>/!oD C!h~u瓲|GAViw7c$ɇd`0nl-F~4]|JfOP˭𽢗[r/ײ0qǓI ޹{L(,~=3CO9=|8Ds٨n|_6:?k"j=箩I__jW^\_?Kfb>&Ceg;sˇѽM`Q鬇:qnZSB B{_:cCi-l{|(ڵݚ@3pls|;Ȟ}ʖW,d ^Axr_ɰm;o;mvCBז>1Oi]yY |QQLa0M XUY[N0g\&Bذ.N $@OHw 6XD E5o8aC2o| l0sY *kP2B85ӑălRovy`}ۊnN%=*z߾U ?J iS⻕v=B}˕"Y~k8APkDƩ"&)2+1k ʄI#bf #™. vY}y9k.>i!ds }5@`}zmy75ѩ3wz }E f+s VXŭ'5Mk A=!`#~!AQ+=֔vH \c1 f'*$513rӔDSd҄\GII$Vd25W/c?y9.C9`3vjf57|s- $ y`La!Oa E!C#z,^δī!J/w,Co;F=z O%V/֠/WME̺X\un6`(g$2%P5J=~MqhgC-EKfHH[sCt~RXCDRfxgg(aߠmpP=qVc^!"x@Dg3 ίoMvvEKrZS<Z"v}.KymaCl7cڈt0M- \w3p]s7Wv}!~Nw89nbl>S*cL (טKC&qI9k4eLzAқuҪ{Cxt/5Zhxu A:-l#$wPTsˑg?_±Z_a۹Wg3M?z|y?xErP`Ay4 bN_y ~5WU >k FR'`F4CqThL3YFq(Wy[@G_ /nvԏfG c2sO?Uw~t3Lc Հ?335J(KnG͆`f<;gϿņ{3.5G7duzKB\xA^ȏNAgZ2Vov~t4{ LH]=,]KՙC.xZ<4h@$S>1fw>jo :!$>PuTg!md\ ϲ4#0F \'TLgD8'cZrRL[QZ[|}[OܳNTe ػ`wf& F!$uCys'Hc$R$RaPގI>ζ_'JFytdoK^SZ)?_5}I7ww*Eρ1<\`:~|ށfpKe>q/B_~gh鷳|͘0BB(˿M}K)7j9g6Ɗ0yL^urMLb.N1QD⩯W|u> `$wׂtKO[;brBd08>vLs m Zu0h!׭i@JO!"Ӈ6Ng.`5`}`v:j%=]U9u؝_y%*CS1)8B$S2#r"0&iJDtb]ƥ2cB;  |E:HMQ']|˽c2 ~!z? X hƸ&{#]}B|\XVr i^]cPʽOL"f!.`KTXzV,)"(O&k7,7R( %J ZD('µVFLXmJ|}`FQuA%A^9M]ؽqҼIaY&\Mgb=ZoSœ6SαaTQac\1v7v<L-'i&dNZ-Ks*M@3-Sk|1Zl2i> AhM3k>n`x;qq'd*A2q‰A"d),2B2!NRg%Zs֙&{ )Oh}5L1g%S\'3Ω`qgI, B8 :I5DM[i{/F**? \ AT/9vQ( )(YOf(<*Y{ Qk ;zLsM& -Nͼ !(t:noJ^lD?V0D*MΤMoaxTy <+W_d{NCW2aR0AΆ0B$z3SMt%[5di}nDaJQE&Gz ujp#Ҋ,L#ɮ0|bcLq5J[D&;Cxݡհl {X*Cgn`T' 3Ը4e [ Ӹ1c pO,Z;7a:k' #zcHbг}cguOPxo5,i"-&D׃"3#q|l0./'ܻTI]JفFtQňt*]V\v/@x;{Rj^u &rn|<?gSL~L6W 59][G竢njNbʼnq>y{6Z 9ymݿ//o@|rwennR?7zd/~ckȊ:N_ś1vEҋonh_cv|+_.@[&&"i^|>YR!?ǣ?/ɅɊT& T_4sE"ɵ^av^)x e/&GY<-{ƣŒ6g7@H\ɝ^Cu鑊'!p(}$8Z @G f[%u Xwڳ&ǂ=IԮ {qt+*՞m`3;Ne&viNFq0,}M5)73K quVloc, XYh4d WhezBx;4JfiԲhpD2 oK,JҌSM%N+̸$qQ`hGAC;?XҙT/<;.0)IyF2kG뛷 u+wzvNњWs%zC+D* Ţ2H F}w*%2g(מ)bȺKB>X\ ?WCSt ëi5D>; 5 ¯ʕ_S/2–JA$0\d֥#Fab*JKJJ`B4QC@# " w"fB9YqA1BqI$SV,Ka.h U3I*Lhtu"`IUnWŴUTV fQV^jx {7=iw8,"sLjmHD! }E tv"h*ƖuYչ9tTD5T* \1ux0Z)M,3eJ dfD$B߱Nu]HcxZ?ǒ2.]Ҹ]O8*IcV8njc)㢣Ɛ5hqE=0-CgrwPt[}y/&vpSt?d/.?ዓ|%u%FEUD˨bUQ(%کkY.6FaϨl#xo]DXI?2Wg-Lz<׸8'Jm-^"<DؽFz ]BB_>+&s"$&R`m Lj#djD[2)MvҬxboZ~(q:Q,5yޙr/ʪl}/XZG>}$ Ie[NY} .?mCBRCgz-222( |UBB&L/}$o=@@ `G&4zc8O7׻ȎMꣂw9۫f],yoW=, b3GlCc8,|hݿE-ޮ8GVܒV%ZcGg&ߌ>lN-iL&`D:#fܭ>nTJ˝w'\De  ¬z -]@3y"˘}Nu4Ho d g&w~xp|ql̋{uMo΃Fӂc9j.MCJGs,xshMod:aE}6'gֻm:1溛)_ŽmMUy(ˢmy?{]fZcA6p6.ߢdoOuZk q>Wϐ 1U,U,N1ΧXn\łĂu՛\ MmrsuԸZ X!I)|+C8B^([>!Q:y^I&szXc^Rcת?v>ڋ1BN;9Yrxt^uCl7{w>mkS.hĝXK<'gϕV_<`$9!t'[?`{< Ea?6נ+&Niģ,NF4\N@zuң= A7 #ۮ߸v8-rQ>V_'`NOQ Y'"&2d5H:!\ke4N8pi\J28K0I{O% օJ3$\*1\P(+ѧ &ʗdӂTUʊ̤E.Cs #eZJXK2V$ .ʵ^cαjD$!jJSӝ*qiS2U4T{J~ 8-'@VJ.2jYr%Zv%|+!'G_к减Z(a.7~@^{c~AjufR2Xa |"$~[$ci$`ԔJesunX5V]TS:ȯtt/4dxa]t0~[8@S;m,y./)=~_£ۑˋhqX0/w{ 7`ܐLӟwt2Gs// ~^@9[ś'.A2kڍS"Av;^ 0ZnnMHdM v A cv}Ǵr?α[0u[E4H;{źv'n<#:}NV݂ ݚ'.!20={v<[(NhhS.LZW5!!O\DdJ? Ynl)[(NhhSoIi`Bj&$䉋zBqu;O~p^j] .XZkϺ`EN"fc8U^XM12 (Joo.Ѷ@H?/@Yĥ,V8qDt"tqWR=gWe*k+s~GC'=?+Ij7붌ڙ6sh^h!p%Br^Zb?O|rWgi6ڴ>ZLoE&s59+s7z6HǝϜ+os-|^AQxa.qpBA݃t YDSދ!0`kq4g) ,R*!VWY`=_?Rhr!-Gl{O)|qbxp@P6{TT^ʶmE &89Sb@>R(6Mm|nO![\4 sViC " A)89ҝxDS0zV/1ɣgOv^ժWH95}/e ((W6sϖv&MK w9K)dz^ Af~:0逦$IP$q ٔggϦu,~$ ^ .z⣹=Mq҉gfhTP_nzPw/O_|x5]aؿy׿a#__v)h)Hs=Lfh8N}G<|7bEosʞ?: ˁ.&;HtA$ץy`THߦ*r8LO8Ok6Qes*J̸U<[tg\9<YkAQ)~Be \줢I&([D:Ĺnm|O0?YགྷˠG' /!-qXPhD%ôtdX9_̲a[1qS͝zxf $ Ev'$NɎ^tU X!@?,&&*DV@0 "JFǞ+dBDv@YʰQkfGwMPmŻb+7w#q32_.bQ~X2|ֶyɵ?}״vSb[]TM9 ^iiHvTUMvy(Pದ:ޏȲ M3 {> aYna0ۨ 1Tfr^˞[k W.=?3i#tǛTUg D[Sf0Iyuo597ҵ齂V]`f* kMxNC_~peGm1"6l`bL:D:aZw75ǀY V |: }O3lHB~&SpιgD5溻rF~@y5eVkAQM̈́`6_uQ!_EH]݋MJjٌ#N/ͯm7_):8UML4:ʜrxh$pQh^AY)-}5='Hk%%{]9ʖutuk`g]%(ʭGU]uqqܒs+e6k?g՘ZX#Y)8[X)%ޓxaLޑwV/NQ`/ wlt)Z(Y{{]>PV`g2@ ^JPPFcGA# A[nLK%ƒ6a,&3L-3ÔD&\l1v 'l{=$&JSRD[x8\0aTgTvu\/;I.TSZn[v&>dɶwl{%5)mc e3zƺjh o`5OF8sGk }X'Jѧчl9M+8#v7dW;,ߡ)YE/N<'0qP ;05̾20E:0d kp/T^%J$ @vD%li!=YGf)֘uiLXԩ5*XU>2x$z]9,(y׏#sؠ#0iќJEg4L-_nüM;n&bQc? Hj ] MW1lc ڞˠ{+/Dl $h1bL)@`h&}6wyɼ{ r=9׳|rh6%A>RqOox5Y S):Nݟ09Ku_A'Ӯ'}<k}4I |Em^]fv4hIo3;7OvKʂ/(m@{%f2AK蓓׋) r\,''ڙ:.)'uFy V%T &yA|~-٠ہ~.6] [2 H3{- zARRллJ&:.A^8ؒtEft]mӵX+ո6_k#I1z۵bܸ6urm+_RoG7iR闰ry\bDr!:DP'#`JISCNdxK[޻|+u*zzzeWԬS/tJXZ(_ݷLV0l?d.|ȩ>ud[NJ)*+rJHkic^ngmuRUOeKT@P0H=j}LP;/GZVƿL/z0n;j؋Jl-35b%%rnt[F#xWL=QWrୟSO FȿKYriP&3 +.Rq%$k]яywnʈu\^[qɯ 㚥 md쥎K)D1 +;AhP3G .MFli,M_P`[磆Lh@5GRbar(6j7"xbBQo8`J` RQ(Æ2I)Z }3D7F|ʒ%jGCo Zs: Ll%!oJ8X nPKl#1HGK5 +j^™H~dbM4‘8CLs c!y?9.cvXc6Hc%b{e63D!0{͂V06/ z! }#+Zr9ph5d,Yydȃ" "J4^Jmɀb 8-FI9 Q"ƭiCb,s K 2A X r}`(H 59RSl!LHFbДY9;A(*UxC v64 EМSFd&0cD!e!x0 q!ce4zl"Po!-멥MK+`sMnbyK7$$$cNR f'`fBQk0X jj| $xĔ|kZ0[YJ)AlHv a`0!L3R(e].!NGi'  bF)yʵ"F`FܧA`&9` m ZKE) ILkII"EF&7΄(9`Q. Fj h5 %߁ r8ZӖ8y-NX5x20ؚ; MY jYGphI,ցEAP`HB0( Rޕqeٿ"Cgq;pi #xMIʎcϹUI-%bc 6YrέTXZ$YVLdBҳ**3tP萘/hmVLȝ2(Wr+l,  6Dy\/&)/uZј2!VوE8pdaIJJP3tc$ OaЧabYJ(dR%d GSU.`ʀR^aDVV` ՠUf ,KJJ"NL[~m \cdN9D$]2Ї dJB 7]>Vn,,0WK +S )" "{s;1g*CI #2*#P sYf4YD@2Kָa%:B4z[#< 8|8hAu~Ag`HH|)kBSK@6VYx3Qv2B M hjDPpzi+KP~Ʌ,>28DP!O ec@.naQ@xJ2 <^GoQJ'hZw -" t lzWPBsj#*l0IeS@gZ9 ,۬[X[ |"KO-HxSW0@Xhr@ 8 ‚`UYB+E{遣3rAnl 3Qb:3V-u0cl@i(PHx^,.zFO( rWE D#OaP@ b D`n`P mQTaF8bR3&hB7.ujl[Jb]LO#+\/kEvGJ!e u@2N0- lg;VD._Ȫ[-kjh,"A*VI|(pb@jJ{~Iɔd>-r &Wd= l`at 'naU Ur `&?8;rZV2"hT. U 2VDT]@$6X/xҁ,Da` *ۂC l{%@$ Y_R=ON[iVa<ӛCa`$ a qݮpd*,hb⮆DEQ4!F84`=`Ӝ(.-E T h,%—Ut FYusIԘ p0!xYBGpm(e v0 Zi :а (N6G.a6uQ$+0-@w:0* "Td鮊*l#`: g $@!?Aj1 & 2Zj ~ r Ü#́;UnЉTIXs LkTpUPg!L WRpxAmJUΪrM3+{уEXY; A$Lzt`_5iINYBe͠%kJmZi̭ |4(# Ox@bFtRV1"J8P`BMW-d0L&Sf3a0 .|Qp# h.PkH繨)dD+UFs7Qni{3wTwS^"o* 9pyВuvA)R(.2PP<*tGdi'O H@v߬GdXS46I#*dO("$U.B-JBJQDF|R8z ),`~  ڰ(RV+NFp `Niu)<72ՍZzz@pkOBA I d4ƥDHc 퀤d wkH=Z#@KSˌQB8`na:X:dС 5!VAX::S"p#FDa8O?}Q36usfGqXq;4?B9ICsCuxޔܕvp{AYem,kكU"VZ]ʲZƧ7 +ZmO&!dqQ\l}qѽ\&DgS ʱ jC8gZ@o*FD@ޘ)O *a@X-@Oz@YӪ`Y_MmTY*GmYe*O0<қf ).l>sd4]hke+T(A - 4,셹lT3ZIp\.h Q̋n,0x6F]N*G i|hE[y{ks3 =_ˌ *2"\TXRcLы%e4z1b6=_ .Rٟqz؝71¤,44Lj+(LP5+逕gzi95}܈z΃_!j'| ݚ缆ݲQHsOE,=rfYWnq~^q2#Cpm77W\Nauk7Irc鴭7͗vKMZ6OhӛMn"@QUy;}WVRm{Hm0^B퉁**WxN5F+S4lUiUtFP2c׿v[ Ź3A 8=YcI?_۷is<%lY(LSv<>qۄ6FeNҐd[5wE#F6AKnpzBX{bHz!c\4E-?vΔ#5q3 [5ehUf5MIm̌'WYZ3v巓vP h4x|{uPNM@\X L4r/;q.X9fobҢ1ZƚmS !?:ň^~|`)ȔP9Tbx)kY^x{r|_>totQa,٧W埧Y(bs(D0zK=;۽.W't׏[iVy10څbOcshlz4 'ٯ~?N^t K`וtq/DuZPNfe. $dFq+y?Ɇi6G_և#ڦi~ގ&͓iVxٱX6 6K9eOOt=%GG ppP_l~$Ù_ hR˛۶?{>YmGhp5gqX{WbugCmid<_={7豶; q ϧo/&M^ZquIlZ Yr#pmo#]a绞{;G׳.\GRX_.V۩R{^TӠkQ.sm+{yR47&7VU`I%#h5  gRqs]NSZ线~K͙NfrΏVjcQ݇r=/gNrC9로ur~?嬇rC9krV'VwݩQ;=ovWZܵi /|ievCP$XGݳhMMnGeh4 kbdʥX)=*ޘh E`Z #ۄ7+nPsn36PF-x\li1#]n>m\UMﮅR8~;n4foݭ[`wR|`f( wͭ;I { bN3tc~W5WN:yx*4ŋ nؑqCG\l1Gϩy+{4/C^Э[ՑJzw7(@Gc16m>PLVV`6BV@-%Rf$ehC(řI Zz5BkDǔ}8LoGV(aT/I8{7sTt~"VVmn\鿄/V\Qt1Œ5&Ì5FZgkM /T^/-÷iF:g4"\'/_Q݇;bg60}bګG|p1n=mzF8S>=9s~Ǐ6cL '+i.l5k?:8o92`Q-vhʳM9(޼2r[lIn870kn<NxM߬\@%6֞$J(q aw''u)r.P:=e&| ᔊQjm_yR}3xأIJo{Iq#[10q8\xBH }c=)H AJ5 :y3Oip 4ta+diR gqܞ)B:澳]w6vV4Z5M3`bC4J1jcG!z4Rb4`JM#>3U' eM"\pd0΋oVFpaۣ~_F`\~5&5 0'm;[`xp/8]d@- Cd~}dgNb19H}63dIJ@\o& 81|Oq$Ee24NIEq2:!͘i #zҜ+D<Qf$l҅ }P  46HR m;=1ClTZPHƫ \`ׂ = vOE 7N" (ė%HS?xJJG`}8Y=9-;{\2Y`dqf&̓j}ɉQ{R(IO7jGDkU>/-EDoNxj]7=C_Lʙ+/4钄QGi5q2 J:̂дU0ȹBtiJ-,hczgӬpζE_I's%5hX,)=KfnOTcoy\v_ j)dA(C'[5|;eė',[5|;~RgW[麶R$j+GJD gRqŌ*I=J=^i>gxM{GڗSyQAwtl͝=Jլ*`ȏoq鸰OV=i{ONZU8b`p?A8]w\):*M‚owSC/l65atn%THaZxRE@->SV/4#D^r;>Y J]SRK 9V>=9=mY@5;N{!A|AZNo`%Jݚ oD3t{ڔNgk|y+ ܂X!e3!RJ8faMgb1+c{Y{۳x%C~3`&]+4@%~ Dx3& vVS3!}9Cy`ɹ\1?gqhZC%&sjքRG%%ժ5)Fau/)WR2uv9 jF=eՕ6^Rb3بXu0ex5?*R,gzN1}%u>/6 !f:) cF< -O9`G+AQ:e<3qA-GI:7SfDLsi[b1/45+n%lLeY g6=\-_o  kOkyJ.؝Vwwhb1`x9-k{DyI; dXdc5!l-I|D=t%klyDj"Ee42=7:]>A jBCH_JE}Wd` ڙS}J&šי|j}ZCHp+e -i=J}G@X^0"D;/8q~4As{{Q&,7VQԷߴ;ŌSGK!(?; ;SzדJwL ±Ft?)o[ɮ[),i<^J-4a=T(L5hc}gTK#[+ ԕ`#.nܳ,iuRRLݧt棒lDB` }J0GPt)Z'B)x,cp-+-= )N|~\ei$4Mtc) 6IK|hB#.UTp(<0!*HVr+5ҁB PT5r9_' S*iĞRӋd0U~lE{ Y5I1g(6-²+v[{Bs,# Kw!'sng 3{jGC7.$nタ 1I^oO7,TiUǾn6B1=/Wbqb"3cS˝oyWU(d/%L]ql3 YH|='0%a-Ӣ2z_](gh2uHa+& nɗqî ])St;,^5]q݇E]OnA\Ps!1YH0~m;&>wnQӱ`AΛ=TRoe_Vo)7= bޓv!sdTVIK""@$-p[4nj%o!wT}xORI$ۃۿ̇Bf0 כOc 8UB(C!͈VԸ Y&1p¤G`I5 lşGhc1Xeg֫]yX3J!Ne硌 X`=6Ԙ"1&Ǭ'|E[փlHD;╬2lzh9X`~|ևP݂=ANBCu43.IL,~–1/2&7ٿw*wXS$8n C7g:'Xb_d]o𨚋LS#P#HFpx4oъV ʸ YrXU %a y fj|۵b>(XW!1/$j3ŌO< nﷵ%y*mU¡1"#$8oP~ !$7V`vi8IFtQ,1IiID##:u2f _}NқfJ3]GU~+3]h/{p skLqǨVgWiĸꯓYG 8m vt+2zhORy5}>!!4!CUEFcH5W,hV|pn[]׀ɴ9尢(偏 b XRRQȚ{ZILfI`Z~6~tuP5M)8(TE 5 v)\Φ}zEnmpc%|枦}u[og i%}aVN80F= pU֤ā)9zڙh=r&0cx̓݊ }ʙdmJHhO*ڡ",nKT?8FDuҹ52mqD<. jP)껇F, Yçm[b-a|D8#wZN;:fI^F fWp:X`[cSHw[L4:rݥۣ/ݢ!|{7ә!$zxxIt0Jc?&͐ A<iANk#۞%P ^I4!5Pl5.JGew.{J^pB[LTz#xzҀ%s2_<%{5kPY&0L>Ǩ).N{d"3k[8XA2sc2[-No?.67Ol'^GYd!r 7|?gI7N/Uw< :dm_#B9y^|_Po_իٝ-t:N֛u_k|ƄjO"֯rq_XsbL9|7rɫHQ:y-䂮3[ՙ\Ҙ D(N6&NзD_e#7d9( o-A l7ė#'u2>>v*J/r>ܻ!+:dU䃽;'PRv[gV!VZ{ܠHW:#X{zQrXڪ9]rƒ~l:5 kY[IY|Ijj $)l+^ ^d@* \L===<БI11>sR3'wWy@Ch{37C9p]YK`|fyґy^{cf. ʭuoCAo p+|fE>24lk<F"[j : ӺI4J)g"|%U8X=qX&qoۄ`CXPP\ECpiCΏET(g;PYi D=qt"^ ŷ#)¤ Ҩ㹠d>Z݃~sUFjnۉ{y\݊juUJrk1X$&p:C cә˙_H/v5Y2F@%=t/I_k.Kq)Hɳz/>;=\h+Yvn:Z@Wq:@@zTŭwX[ vsdNUFBY`j8zQ G=;VN@dl,uzHJdo2&abg CJκOIWhpTzU>0NߝB?}4J0[yS'%gOot(N J8vB,"nL퀸ТƛM`Lz hVgSgTWc-9YM&<8nn.UDlR Mj]_HܜLEN%;`͚)PMvdX^;Xà<Q5hnht!ֹL vNo1V(b>bsxa%p&HW"ƋR۟q~|Sxi\Hzsb}fV4.9hǒB@q3}ۿovB>B dӲ2H'6Ofx#trB*%L6LBwSU>+.WN\9rUQ/,vu0cV#!JBC>M"F&EbKӧ՝r{uxytFm'+JUvfj*֘:d_At'7w]|*Yvzg"  Szٲ;Nx)thɊ,,tU;͔G\Qlq+5s>$ |빅Tƥ`8g?;;ԧk/9 xe;}qM2j^'?э;,t±F~UyyrϾߖ: r^0?O+XZ_(W8|c2@h/qFXDd" Xsp(1r҄{68,D۟O8VJ(@nH J|_ǚGP<nZ&.kAhGDTد1#TjQ&xjAl*I4LSOGQ7 K `mT~{k^[pu\V7}/S8r­v[;;ˁysOHcxL//waA`vQck3i$1'bfREZ }(FˆiPJϩhD Oy}Úrÿ8ׂ?^9ryUϢ+&X2IDKAYH) d "D!ؔ~mnK24?+*udu/GUQk^G2V}}*+ ShNL'l\U2<Χ0cHd hiEvpUof __l ^%S*ϾHn@!Pgן Mr]\O/_23UA _SwШ^Ξ2ACg܁W'S ㉹N5(.~< [f:K>e7pѹi_9mzsSxoo6- Ŵx>_-w_O2Oayuxb3o;h/}ʴfG6):P˾o2n  W{2/%j滵(Ո-WeRhIA ܚJ_n,?qCv%ʪ D)MwchI}Ą&S//i\,Wv$%YH |#1-ǃipVєڑ_|/DPR1xssYLNX_\m4 tWpr pctVЬhclĈLk}-b;BdI\(htU] (n!ן?wq6NKFpy8v̜j{%O?(ꀂ3pj Eò鴱4;|j&Rg5@qV uXx h[Ŭ:>8X Ν&xPj{tfbӶs48P{,DZ3`t¬Y$!MLq0TGI㕅F!RR u@ t+9'(.L+wub/q+ǵWE-a pj,՘D &i ca% #+C) D.GjuunK5c"q!Hjl~D+s\h#4T{=zggcP) w9d z {;zZsěS4%/ tp%@:?(]-@]a&%=sufbV{=D ;suo2,Ipq8:Qw+ #A>Xj;+\I)p'b@w𩳅~\KmY5G܊OJ'4.u|w׷>`qx{qQ 믮^|}RSWo}51s?["|')A&gX~OGfuv>cɔYi͛^P453f9oQu׋J^s Gpu 獲y#9E]G\:d2 rbwJ)J!9V8'+]{#o>h\E'uf 0>KIz|z]"A 6yFpGŧ;)HU9)(7SL JSD]JR`dJ5Dў;B%%ݥZ@)eWŻNʅw 'hY`ĺN@%T9&C!ubŤ@E$hNXa{G%ݱ o h QWn epm*}2ITz@x奭%Sn}J5?bQOEsGAM)Zԟ%{+(K(@c \hY P9Qi"KxJ'(VJ@*S`  &X 79Z5m+%!0i~K3 W@m*bW1\SKcB)v%X1I kU`#+hwOʑuًvg{C1H*QiX[Icj*;<l \ 3X㰳fOuV૿W÷GmX],wn?㱹ƧO_mp W-\^-]tk[AOs 4_&u< VN4 02JQ.D*q` b=q {^`6T)viZ5^ƧKSu(s7f 9"KMVSkǭTXƣx5E>:fztw#SDMFT#_<a П 71YX8FWJGC([~QOV>zRJz*'yXIžBG UܜBܿJFzNNki#]F^C)*+_KN< D練3V!u̖)4a{ D333+:g< TѡF6 ?59BK &T7BNkל&2Z -pM)fo k76vK FtREMg!6u9R吐.I2IaOQf6IF6.R>b jEL1dO5]]Q2]mW"Ǻ_ZK?ԡ4+-U5@ xtqOsyCK:g+ ڹI}xwTTi>Sh4 ria8D-]/y oq[pU,?jkn UKsZ}ƙ5f8m_ ^YN߼ V(kdݼz޽α[Ҵ*k0]8 !{zM taT w՞7."U{AɾxA-">A_(:Wql#P RBTD+puVby/{UiZrcPIZ)Nd@J  mzWwZDdÔ94.M CZ#ߑtfۜAP'ɗy1?dXO",qV 懑>1VZ2λ@y+ $* &{nF@O?dk!nSr~8璘$9$䅋hoFF1:A|cnq:U޶|^,CwR(b77A$hR*q:1@#N'^N7~~X-qy| u71O|dָs3S:wvyc}1њaĤ"A.)Qˆ]|"Ib 2D Ax+T,J ˜&Xˬ [v ~s-dBUDAR@AaVQ ZY*CDeW,)d :֪QY$u[sA)Ui \!|f"U+aZzJ8]:s =eA$TeYׯΛxe-$ JVD+IhB6Ā- 5أMʻ3ğǛ.>|LK{FJ!/?㽭H3IAxeF寃N?}wa.?_]N4}J8mU%qO7"%O p1J@RH$ iE; m0}u{JtȆI?RF>%)ɹ($zf3K($Hg*KDil8HF )(( ] hD~uVӒʤO!֝ 0oM ɣz%A|NDN@*8'O:JZQg C e !Nt/?$P!`{Ei~cu^?T+q=(LF*m4@V:G=҃%~URA#?J%*34y.Bc}@ Q;WVk^_#kb2,o]7`(,Z}Axk8Ν*}0c `Ƙ/U&t<>|pe77O3)xu'a>H_l( W#pe<%"J7cu4r\vY N oM}Qq%rE%@ifҾ% ˻!S,wHi$uFϔLlWqggr#K>(xvG{9`=Z\JL* SP dEUBj*Y X FYY=^g 났$=aP>EwIwnϢ߅VfKHMExLM%jTeRUԥ!PRkN%V, Y&h[,4]JT # %C CFU u \X'J!ʁʁ Hb11JF3E` bpLL*2P$/% ;j{%qx\Y,i+hҬ-~9٩Kn&o^i -õj2Iĉ|p8'#^K[y܏A(u nRfSo }?N$UYQ뒬є9,N@9HFAwL&{ 8>=nFC{ܤ.ޕĭWnBxI9/sNDd9r*(JZ!4 wRw2_"0nh9cJzgAsb=:$גpqTz- E}D %~S}8ݗDx/*bf.I2,Ln>햊A~v}Ilz ݒrH $2tv೰R1otnS"nn9$䅋hL S%_M5{)ntr*1.:yJI\N9$䅋hL)O%5B ; RpJx zh >Mq ͐Ǘ|#7zdba1O7 T `JŸ yI橠+>?َh혧05OY2Cscs 㚧̆@_">,J|c Dg@"8P O7HTTH(ԡTiƬ$i^1b4mGioYWχ|w8xo#`ry;Hs9T9Is*y@22Bm>>u0`>]E>nƜu3SfVItW,mm) V) ո4Y*weK&lE-n؊j>$ d@&: NoKd0A5RvPp$Ϫzi'مhAr { _͌9$8)${ I@%9(,45q8pa f,z<5'zL-jy^Z\ݺ5λUtcRKf=vJUmYB3Q `3Y>&XQĄ"1&e5C&wիWJT Nʡ]8Je"c" OQsiTFH56ƀ5KOV<{$ J ja@HضKQa)42Px"Y QxZ" 0 f> R KMBJQG`<)X{]:Op푱iB+G)" c=Ym#4,l8#?7„I-̍h>i0C.7HwA5?*1FQ۷jyTʐyUgc$-G' =B2+Ͼ<[o5eΚ27i!#)}k5^~evy=o8a5#锵7?L`BSLY0D$ fTE(F8zf(G`ݎUZ\5U޽zd,4mmt Q0I.eVm 3M'3vW0O [IyziBlӯM{3v54Thm*"F4pa(NNpNSz!pU\8!].★?kn7~wЕKSwJ}vGEQ]* BVtwRJ9X\,D[נ>`Lҽ]lp5F .xZ,m;WI?BA?\eSarc#E»䏕4|qr ϻ{>yy@ɻO* <5#_c/3BSEȗJ'}#ԁzS&`Ysl,[ZjEE&rF( چ4Svxf4:MxqRJ]YC"zDåX J]MPL= ʶɍ21h 35j˼N5g'K{.Ũ}gѝ?-ŜfKjy3h羴S=ß](rf rrwL~t AJem'FHOvz={MDsHc6޽}z҈h<^t Zgf4댮?ӟM'WMM̞n\_QzF0A NNi/K9Ҋ1)Hs+HxBj!0jOGs)o(FǶpm{dz  9@arJ{e!S;N>x rEO1/"zNkNkZFE4Ga2il(b ђYg2`\`d()"* %y.uEa;.!4e#_7ZjOLS wPAjHņ!!yV3∪"2!`'3~U Bfe_i)ELoUh4\Z˜2hRsĹ&`Ĩlj<8:*A 1&*d Xfxa6 DKH o#r΄!V{/MF c{; 2YI="\jHKR0Pi(A;@Ud78hd eN胄 d(ҴG?].ʛKea©HTA)M[uZ4̓NV@ArԎBB*zPizYqb0exDSy)ZGBN?+E^ /ŜU{W $>R,Z]`\]_[@?ˏTQBaAFA;xWp7Vë KiV>i0}#zRZO9Ct4Ҹp灗\D9ZVYI3zIwEd$7 t j"Jl^9 <)3pMzwD1Py2Srʭּ3PrY<8>JVy0tDò[j!´' M@09>0 D" ƪpܱ,G)=t+YpkӧivJSZ*bZXNw:(G"8_0UƱ#>bDc/]^Ejӫ|ki撣s5sa7"?@mzPO="Fe[:4NKUzIF{ARK660lD&1R殺u]RsCUjV#-R=pU^IAB'2\vhd4NNs;9\O\X,C63ŕQG rSH±"5J'U_T?%gDYߊrԍ-u%/dSM#fRs8\h2<>4.zz FKvbʗ8LFq078~e20Zq-j<ZT}Gt^\ {teA1e2QWa7U@A߅h0:gP6IkҬC4ko 6d'rиGI+9 xy:D&] "Ǥ(2'b\fk*:ͤ-Og$Y?>^Z7y$(I"Aq>\ɿ}*L_γ랆> 3lχY҅N>Uf__*YG󙪉u\L=2_Ӷ^u_\76 ;M`v||[34&uzd<%F~Z;Oo_@Tgo'/^|,^ ;\Ew 1ʱ= ~N) ˯o~θ1P [l'nUGtszKW6?Av;eɾJiT w΄O'x1pd"=Ko_@J"}+83e/S)~M֚(1CWzғ2&9/htzYvqNJoev\=fz-[3~qks]eШе/ =(Q9m m6Y$?bG~Ug}*4{dj?r82۟Kү5\4 ꠸XU5)c: "zii,@ᰔ؄XW%/Xf5k1/1 (žG6m%ՌIA۠Gg Ϛ#7#md̼"ܝI$Cp!GΪЏy<]y0tML8zYcq(ʥ7q-4|LA)˅Kip[*'T^~~riӍH )kϾEpdOW wj̲$3Ac) v1)VB4(.նz|#D(RVV`yhI7 #U"z®>k"(M0 ?Eqc)3tp̙ƤtOTMrP <K&># #E4F%䌴g"#%"!/}n\ܗ̀Lnd܂~ӻ ;wVՂdy=lOșP o p33]J;IM0]4\ӛp5WVnw5zt Rp)j4lJd߿npt'~L@+bgI6+(~&@sڲn6. ,nhRq_ EJuL\I3w̯|:r 0>WmfpQ"$8]Lt^\fo(.lD\kNeBZM)øinhV4ng9}s7vq3YV`ft tbTL!e9IgVgDSrF@"_=!݁JQ_uB:W_gu_gu_gu_su0:s=BQpB!dCnVBF6B62I+V‚2&IQNP\7<3C Q;ԁB{޾VH wп/A޾k%*HB}=(mqo ȍ!PvyA]$n[DʎjRr5fEJUx6r4B,SVVbMa22khc+K3ihAFB{$TFT}'NpM&ܮB;L8[/aM!"l#f71plBt&j˽_ph/-C-A zmsmeH{A\ D ڤG/@8UDž?wH(D )ύb\k"(r5\-pC)VJ:kglzªGOngÓ3EsɱRpb]I[C6&Lg) hbحby*jgӉkzBϜ`Q,"eD]+A"Y?D "(sn\Jf: T}=gb7Y/pI_ADs*ablRswO&!XgW6}iWa߶A7M-x&S;@N5z$,UI3~J:?Thkы463W`-/ŌMUKďYR.oOj D$/;-Z5KtM pn@8sҕ6Mk2뀸wL~qQrPa*(%:>T ]M4\Ji:=Ta!6PgF\26<΀2GB\qg9,Y},WK ܋O΂Fhj/J;hx)MqqYҚ Z:68KbuXp3n$ZG.h$U#^6,k6M85ZU%IkçW$䶑,@? [}^:tyIx[$ RriU`s BR06"7T 6A?+yͶDQ嵏bp;3G_,9u7Mw{4NTZصѓ}Q><۾(WР0|5ld 8 :}S d -ӧqMg(>Iej&. }2z;vj?/=,"R>0sTs [yҞ;kl7Λry(Kn}-k!$ӂa‘Qy@,>C {1%M1JI|:uW~PZ9K~ OH TP^tĂJHP^a((<$hCnap]sMx& M8hA_xj­.- 3ׄ4heheaw)2PA3Sƒi3j s*o(uL PLq`݁z^ˇYW5zc ʸ2k2GG_gt~",_UeJ?a2a#b9$=Z;XZ|@,X,XտNijuPE`=s h3 (n* mBqP& mBqXܦ ֡M(nsF߂$劆X;V[+Cwnu#t FP7BwAݱvc'q% q`3HAѭG﬷!Y:9F~et 0!.s1J:3!|*7|BcBēIIA`B0!y HYă5d&t,P#e Q퉊J<0< 'RpUH;yTi'-1,=wH;ٓvvNBIH;;vҐvN<5aVtʆ:qD5c:]kw5bk72vD'F]7rQn?VuT& uwFg7ߎa bm>WR+NWoG)YwL6_+"Y8mހ(lf>^t2Y [$pMʛxaߺ.̾"(L:bbA0,5x =<zz¼Y(5.O n2z49&h3*ŴIATaSZ!-\gBh$qwAnyY0ǽj`6LwLhcg^. 2(p`ewKAegaFP}^:tyIfJRNQZ+m3eyfw$KSVTjj5KiT2QD]vܟ>^mӇ K[gQ7.XkANwt[\>}GvQO\*JpxTAm19-hC?=RLA!h`R4sߠ{W@ l4Jtw(xR I7q =:K<rz,eTZ@)/c$0?j-Qơq(ʃjġ ߽y'[©xpo>s&֞.0(x-1"Y46`-1;ؼCK'@ihZb܎j"/!)ްj?~Z*C0Ϛ۬X1 ~CT*!sD%EePlȣ-UyBǗ!$IHnә۴90x.ۡ.[晸ʊ{R/0! 0yB ?a4e3AA1،iG(m)KNӊc[ '|ۛ,g=N󏗎[Y[}l@VMUQ}zN0VByr^? ӊňgO?6/5F^Wg3Ry_" Պ M(?9v^tB AXt r:M5<&qŐfrI{s=i它F[+6W=g:*`}~7oqosf#@@0g<*&r<$UxYf7*J1UjsЏ`[9FU1{64}6k_xkq" i܂pM/]BΫyh!+x 2NwqhIzrn<݌Qotr#(U[V;Zc8 2z;]äW9R.?8Of[K lqi@W\^~^^8 F$Y'wesQ]5 9NXV]Wzյ)|)ڭj^v`[߼Z?NOxMpzwz2U"8=sz56Ynz5%n.>[o8ݼ_EHp墽 BRG<)#$B#4S )m8daju$&YgD~4>g`w}(b z5}txȭ٣[5&tVr% pp],VDcm; lځd7Mݎa*<&K2 -še߮?;!AMSAA7+@/^]F¥<ϻjvFg[NaL~vaf.F(/ſQehtMNow6AIkf-V礬OϐCw܄q#ۂ*fx+O=k8}fstr=1W C΋ڄ|mjt1 &6~;׆0'.Vh# &!/t@Oo==b7TKn.Хl4(:%IΈ&T3zn7s@b6pW}(Q&e7{[\0JKd &McE\fB -Ӝ#M^h^]'ԋ;ԁ=IŪJL{L@ۮVj5LiTQ)4JP:pM1L3nZ–aۑjox–$N NlNW:= %7̝҇Oh4XzgPQu3D _.NaTr'ʰfn3jn:kʤ]Lv!94f3Ѷ=h۞.U#ƙtdz,|jѸ䳷(/;tg?t&-0trАH^KK0O1Q۷`.!Jt]͝A / !Ai%~=C% aMOVeŋRWmf>о!ce^yGü7y5tmF?X||ediQ>#4v]y W}=@F44>a$Mz .u,Gx)HLq5G"_U҃c}AuUQ1S|?ԅ@D[+HK$ cم@DģZQ :s}J3yJI^BO9ģ #L#^sE88 vhF͐8h#scg]`@tQ 0}ިtj,%ad1 NԄjd9#J\ mZ B9j1ِ{ U=E{S\^4qn78E,cj)+"Z$[yxL4}t TU?/_s%JZ ʁĵ%M.0j+3']bRefk:$E|O*O)r֜Ѡƍ]lonm V*~"nTRV2v} \s!о5`"tlyNߊ9 !D@}2)08\wW}ܟx ʙjStČҮi,1P+u@th#%!Ƣc66p>K!X[&&ϋȟ?/"^TH(h+.1BFfq5N.!d7˛ 6_>}0-ӧi(ONq Dl|r򿳓?73\Llvv4dcq'ϗ0q5e97C'?o?m:?5pb|7fwo|i}0tݗ|#pϿ67YOտ JЩq_ͮ?/x8L.N/ζSzx퇼ժͭ,~xmtYe9%t{—5_~Y;h; WD+S?P ?VEa1 U<zu箶H˓ŖkYpML}#|0otq%6y!8rÚ/Xvw4S<-!CNS?yV0S ]2t<$U*pqjx&3IH$8E. ²Hpm$1>FH  ?}# x #a9F؈>FeDeQxcλ΢2LAX+!N"L CrSV>dP8٠aVLiF,(γfP$o4ku(?iW&oM a#t+Y.nMRo:O>lMtU u.IA#W77?gP.yvMxQR[;\Er8f ^AB+»tZRMhwC@(5.wϿN#cͼ!c"p L捣0ʰ.Ʋ3)VΨ:D |`pw(g0myrr!eenHQKm-H+nOQj)TgR]ǘRUG!k*d8Ƹ'x4WJnIsTvMfXU6QvG9$Dr1IRHjScA8`" &Gz5⒌tq0}TkV̨҇f F.ǡ2&Ũ.lE9\DsQ9)baX*'1+er! *'3&b@gVύ z9iG.7٬̡.YtʠgX"Ɗ1VE5V25#FA-(9"RQr,LQ+HpCkM;jM;6S .0m\5&e3A@aEJ{nW$hF_6Fi0p $%Ť;$Q[X&Ƥ`ٔ'a+w  ^ MM );ȌG>;8Iy"S!)W+RЙVj(0RG:c$j *`ZWOf z:Ej1bkbEڲS1 $n;:XKݜ%}ҪH(`9yJ)e::- l3b+N@BWiKEJ+yi8Qo(4%#2zDC4weAHVUl yG}v2)7WWNV70ޞr9u6ݏΈ9hlLȴ)/@+!6qJ;y^DaNT叡[)}JTJ+W@]NH,PΨJVpOUa8]Gb@Vk*fBں6)Bz҄8}r~-_{h ={` ޡ@k^YՁy=Vc9gLŇ_@0_(\$ y2fTb4! 9:+a0t V,rQr<1!GmBȎ 9 9 9 9OILq!SSr-FN1{ 'J{廧 'Ͳq [44[ JԚ g {Z[[fr a@7Ӛw>7X!`Y-FΞs[!.F||  XcOcΜ%,9H>4S>*,jgyg JQFrH E !@xciV`R?60,muPhc[{L"^ ;šH|45Edr+0IhbtU.vY"Cv9⌞Bw &˰|j3tUCm6(c|2SeBZ*|,Y~q^D懬?D-oX4^ꦥ.aT%= pP%*ϜB I*"IjP/?6#HtyjQ\JTb5J>aVQYbx(V_"'|FaRI eRGpHo u4BCyW3LtM#L & SnžF# 7G ql|t UF|3XşrVLiCϬ9q*L=Y!!LQ!_ل.hj#h)+-_aD_لЖ:Lb Mk۰SHlwW%nfZC=~6_?Qs]2`wA\a6].{uXf4Y Ys7&M ƈ ps^>q{sv}O&j菓 LOfxm6u h|aez[RhSFV'*ߴ?;c6ɛIMٌr\u&;mz6$:c‚ .ײhv:6ރ_kn]fӼ#ǧBNN FNճa;-b%eW5- [0oͦG f#3[¿׆n׻) VjmjFxZtz<ԽAdO˲Zr$Q&ATuK!ZT(BZIX+)D"+p~iՠ8of(<+kŒk#0m F-;C7^]Bi#z),ObKk#Na 1 =^ C1ءNaiN!XPv]t lC:CvxN-1zUp4hܻoBTBCLt)!|G:{}~-QiyOʪFGm陚8@S3YB.-5s8_eGݿ3_t'[k~i:1n*Trga 4ijm杬lCYlC(!;UI:HӋ$GQkIۛEEd%4zQGKh9;7WO:I}/[5?i:EzP,xNw,!iN%lg@H8߮M9~L'4Y,SWr$oxx86T߉jyQz>?CPɂy1Qs,BG6*`}وYΟWmʦ4i}`/oO}]zڥvN{ ҽ7]D` fTp-AXbR+Fq5Q)=7#3/k1ɁWݬ$Nl~Bm'0Uח;כzH0<ٯo{3I{KolB"K/4`y}$ڲ{*QFPJ!G# 8u1o:R,VWF{.Y rab̸>Ma4SAX,h@!? 'n'X5±fyX}zc2C>-Е^fz`WW=X\#߃e%m k ZCjFlCk"vV܏6E\C9:Y.bGxT BvOI҈Jʳ#H/KwGd."p["ML}8j L8ӂhH0("`j$R!\ ,=1APG"A[ e"!ƁD=֒q90li`$(zWj 68cM T:D3)i2^hZs5| .Hh XoeG`)z/>V׏Z8^X<^X8^X> O~FhbđŠ++/UoT %~ԩ7y*ӷ…'/F2*Ɗ<ձzwX5}KnvRsF^ňճ}/=\n$XO}|Kbm9\ }!$P k/^lܽ-_]ө}%o =9J/@]X:oL冾8751;r4Lvk@x^]߶[R4Rvl/-b2_%wHW}@I0;i&BuDU%#ijO01B;ɚdGzPv2jHyNOYG7jQFo/iyc"%87s*j^ `-r|&zSXJƓwtg5cĿnL?w vG_Owa;|REv//|͍ =Y4Kq%wXaWÚ_ސ͐:k/5_lp!3[˿\4@Uʾ5{~*(޲}vmGh<;ѯnAq}ndRJ?כ0~\}ziB 5 _&J4lfH=c jZ$, 2kwU>U55!G΢Yc^iDS nMpȑhOiu+VO jQbPurvkohZ}?-uM9f DD!h\ N36n-2'r7_lvkCExJO\cȴdT)LIԍL#g,ObPurvk/ihZk0vkCExJF'ڍB\-:M[[xYP5[r,SDM*bPurvkqДio5[r,S7V>nșfbP+!rmz-cݜir,Sl vcDA}&ڭMQB~֚ڭ 9r)J"F]}koB]k VL{=ЍJvMe9/mR8E6Vr^y[lB@VF%T h%0~V 0+^+=f%Amv^{A;XdKI*do7Ш$P:WM)DcsMJ1+81EYuvQsP]LYsscnRضGB9fs}QIEm1Ǭ19%xJbs ];j1Ǭ&19F%AwD>瘛LtrsJscnT1k19c1OiԹ&Xs@Q9f5s}QIu.Ǭ>Ysǐcs9fMsǐcb{9fʐs}QI8{9f}QIxP939f^s^=EsGc1 UcsMJ u/,8cs/JSD2!'8t<'b_vaGV,FppM瓸*..g3đBy$ +{eq:Q#rwΙgPf~~Z|ӟ/L{, ) ./ H0+Mm|7#tp^Rbf  6`+!r 9,p=8Π' Od+; !N8c[Hb,< [YZ'r@RA+ z5$ڀ !%8q`bDT$iT8gp> Mð/q"9ƀ!_Fc}z+85},f1p<=rN|a!Y|=YnV'hG\| —Wtr_j~a\x|"/(/=yH0aH 0.ʂ`&H(ꏨ k2Qc7 `R p Y ;-"Φwpg#B"fY CbfZ94f\oVq-8a]55tG?k\W :}=K*IiKPZgygSŴ4󃛳 d6FR޿}w{,ScO-^x&Fu9 "@�6Y,W7L-.^\v5~N4LҁwW.p 6׬{2WD7x.;4›p9sϟ5(lU|pmjv.%hUW0]WVW&_JIF\SO`惇gip`)Qd`q݃|/Dҿӻ\~gY4 6"BC] -!Sq36?d` ˫Z2}-jp #[d믉r케*y`KVƨa`F%{'2Xyo{g¶8)K Snm/hZ9ʵ8uQt@P?<|h6@>~ؘ]gc|hϞӣ?"}_#)L…'* +O>/;^=wFߞ䊇>^vvità~ *_ M! Mw{^Y{חYqia>wN~=`4{-s+g0rV.v2BNX@հjЛj4THt4vGã}=UmD*꫉a:l7~p# f5L&~c?g@}Og98 …dcBiWvu`3?|mq{XAo(+~ӻEǰ\ex ? ^oޫ~BR8vC m//~@Mw`JO`Ѷ~ܩryP02&oL"|z]leK0VgmMyEc*~|*̦U)f2yeqU+<9XYidLN..gJPR(`˂~2ڥ@UJ{t0σ(Z}`qX7dqJrz:"EJ ÷b^zO؈WLL.F67 ȴK$~Dr7yZ*ll"]+Vd]/.MviKoKWX1LH3N 1O޸~$ wdK%kaI4R9;ȝs9D1q"d  4q+aDӆϝ`8 8jR`V4)bBC `./,# $yHf1Nɐ$"gLj&2(*%(1`xa- fI@h%3Aİ.&c8uR 21r]`RqJyΨP6Nn1XNla+&Q{B߮"\}%ruzG5ѷe7>@\'_] LՁkï_:y!U"(\iq`<4 Wo_3(~`Mh9LQ'X1lx4E9o_ޙS}o3vNbl}"}3\K(*{Vj74(+\? NJ'742J-hS7UOg`b'/LK{~2c 47Ͳ?|Ú^#kd/IXr /ATp7' qB<# npjAAawi΂>r;;?3w($ܸR)( l|>o{Z s\6S :#Dp0kb&Љi,wְb_ifXI!W˳0pwLO V)N9u/_/}дbǁ,uȚ{myUBR܂"0Fvf$fވÀ"dGH660iی "[^ۋ؃wiy^酽'0>2#Bԃ6ȕB5O槫,5I=2gFx5bvg e^#b "Bgz 1!,5v[cr!VQohc@w F !Z*2V)l2v11q2Wm&Fr357O(⌦zR[ KaS=l]0CR~NMua =jE=l>a][KekzGY;pmacJ,_7-L9L;C4UI]uǸ=sޱu:[i y7pq|zY_E:l56j#[QV;ҲŬ}D1R!7NCd[U1m-! 6AZwMJ+; (M 6 LQss=#yo TZ"xxr6=gS- RlJΦlJΦlJΦoTKePTlJΦ%b n21oXKZoS Ȥڠ$v˞zD;Ƥe\ h%)v{"Eq3q*Lú$M &TviT,[2ɔ4S {vz 20:cKEsg%Y׭t?e}5P1Cn!D Φ XS($OF3GK!L6߀a:KK̥z`c(X>9m("[I} pXZzc!4(Hޖ 9sd.D%ZYARj J`Ho 0 82a܃ٓ%=zkUy#̐F[*'hx/uT!6uF3AӶ;آG)­8zqPu f3J 0-eiߟ_%05g;.:\\-BhEه<.Q/fy3"$# &q]kZaJZ6V{(l~_@͸V1~yȁ/vl\?A5oFAE89,5v~mf7ZKO4i@UL~S09sda!!2{n̤brA,|u2F_e"n޵@Fj0@:.2;)b/S:nd>P=1-E b|EeZ1=o Ǔy8j8I; I$*H J-T㭵ȿy"I8=Ŕ"SZ 'ȏm j_SIuj[O-etM1(7ck!)rY]%F)qGR>&EqrΡY$*X 4I%Ĵ&%$TbJLSi*1M%^iMAQRi*1]wqXХI E1IؕR"3i)9)<30,bxRx&gRx^3l EiE7arqJd2lnxq0 .<2$Ƌף<%#j T3 ZzP"e]) NH[ݨᾖR6Q|{n',9% sHh*eƤDF+e +@\;\Y.8wmvu~m7 H bUe/,wetߵBca0zi᷽͆ϙ7!1& mcQijw @Ή`oAƛ_=x@%–u HX;[_zyF{ӛݳ*mw{ wQ I[u]{mptY.yAKf6 6p$X qrm30@)d a#NP$9Y@#Rq 6O.t5BH8coÙJ2\PO6^&=Lqz2") bG*\]|G%^o„+B5s:ŻO_;洔g%v //Y9l ̈x6MB)o2ApS[IKS~N<{rPD5h@ZJOyA(S9?}50)tTNH- :uTj㓌#hHFN1>],Z8_ :yGL);jꓼ-ɣl(,TȦg!mCUs;nV;ɂZJS̋KT2Z#G;->m>ŧGEhQD".-j)~-3ۡ/ê]L$6Vp"aE:(R*Xhby*tJ1Pו#(č`@SĖ1КؠZpI(J>C[#^M@dbhpI=RJ4gBM-B% 昗q-3G6'Ne,Œ~P43FAcZ@l& DNt8KuBcd<ȹNC(cL1D n֗ȕllqqܗ-ؾl4ԗFcV_;,1 B(H qc:wA+$,NR)`IMAjr^_8NmM)cq \w 4?~ȳbD K^q>0XRTܝXWUSTF!]U*RT?EST?EST?ESTQ4&z8YiT&"eBa"[nL mS i ;EBS͟cް3l]{oG*eCaȇA 8LMX"u"_̐>)ILuUuuuWկ0Nݮ"3p3-u(0EipEW3L[$Xx  xg"Ia%z@0DӪտHtSlݠ2I̙>dd‘Ԅ?Mƴ ~|_7R\,oW3VSp|,^. 2Ύ(F1мs߁h!PN5;|T}Y5th$WDh&2C(2C)F"PkhxMS@.Ёk'/ 1w$?r<{`1)4;/}થTE}Ӟ6Ew5T۔Lc5*o&stv$Iyqc?tG$\ZFXJ M9IHđMcs70eFFX,xuT "R6z7ys Ѭ؇3kZ+˸4X[O.9_s7$sևh6dA=eAGHq,X~K5)2)Hɔ XY$?g vˆձp}-lXIlwM&{!s$]rU=Q]·2}fKqBk\g#F'˻$%^Day7 ' \| s7v2l5 Zq( u9G xk U*`{ҥy ^HwDZ̢mI? m`1m/V$p|bfO̕l:Cc6 1X'|YJ+yδM%k]`E;ɨvN'Ԓ G` ~$Ǽ\@RN(:)t,ƞgޖ~c|&Qfy,6ZhJtʤ9'\ks 6r6|-e1{s?ǹmCE^3XfŎ`B3~k{kyF>Пq^\ Q${ >$ ZwRs3y\X~c3C夢cQ֝tK\qXī; a x}?9*-nt}"ݻwtDðl"݋ @g%Ci2x20,134UeC*Ք`[a 6}Fǻ tJYص&H+%&4O5wt~ID%`2:yϕfq \}jWc鍳1/Vv̲KdGc@tVn ct}L_+Mv& +LH5^7C 2yَhZr`l9>Eǣ ;R}ӂ^76R3đMmdUl{&meRO ZȻ=C30̈́z2: LZbJZúuz6ɸ%yJ.#s-,dZhue3$\ꖛU,;EOSn?xF D~hb6Tޗw e%hµzNa<_.0Yqv(Nd+:@Psb,jeb2. 5'?OR4ev`yħRWj!NVOBa ֤TӤgIۨDa&(IUBpDFXE:rT̍ 3!FfFR"WW"wYC({<ܪG6Xil8R8SeSD(\h39ñ1rgl$lD224(U2CyTU"FmOL5VڠY\8Xsd)P'"NءZW0HZjtC-$RtUC:'b 뭲*Ls!#ldK'I#[>~Ḧ 2&HEiս$qw؊G(L%^W5vYמz (Q'崿$il۸}_K6Cڨm5su|md}4:R`^al#h#Ạdƽ3ibs; ew[ I{"-[;WyνݽKq7-u Ct _5PC۫}9aB7ow ?%~)4DPhD҇Ck O#;_ua H?v*X$2H Q,r`-f0 p݄:Bڐnބj&dRHYxcTk.;1j w@Ci%Edib+x`4B(,R1V, 6E\CIY_ބ\mQQl{Mp˂iĦmh3W:Q]5z˺a1Xon|T_cx/t~}&ܥW|ɞzys~kWfL&_(M?4q/nwk~x;*fbw L+0>pͦ2*#$3Ӆ[՜ɒ3xdp}/Y#)"IU៾ OEo/ \.z3$?21L7cl8҄╟$)24/4fĦLb+#"l"Kj$U<Y Q{瀾yk^{ 1;7g #I*!uyCQ0L{_3 +ޱ<}+r̈́>E$4H}9-œg*T tV`YQ^MVXy&01b@wmI_af7 PeRW]IpI3 E$e;= JC( NU$ 0_w }$,TskJ4pµ)胢VQN6NTpDA |Wy1n֥q%95کv$X=wyπ1Ḓk >b\ >zB|,*1s=* trdr6(<2M)mrҶ=td!)O`)T$ ŝGˆŋ>ZC"8fxaFPWKo D2pXf.R VP:kYm+`֎ xkH#3os$V29r% #9Kk`]=6vjAy[ u8yW[ᨒpt#%, / {#;e֩qB3穗F7TgV p!fJaoҎS2!]*^Lrhdpȷ,. nÝ0E@vf$. 8saStb&)!"fDiO Mh yE 5{4nq3u U+cmzN|t\ٕ$C~앯; o>͡rI;)e ƃ2 f>\ʚǿgeM|dp[`VJ C׫R~G4^$̇uKT %jJI )Ut},,*'PH bhIJH%Bٟ xs-zO$LBH䐳P$%ac 0DwM(KrV!`rtRimj}/g &`v; ́#BhLײs˜ 7 Dpu)4@}+}:O_KICSB7f.&;Hj7"* cEuQU1D',O曌WyԼ<?PDhDkheL" 3&U/X̷WB<;!PO`l. =5Rշ;gpk70攞tf:9穐.+ϣ#N:9Y9J$[㣸 S3M!i]C|(au/r "çI -h #'֡U&LVMyhʼnb"mj6)im6H` L"CjWVDp֭ԘmFqL`U㜄E040$Z׋[.Itfl'A'S_2"R4@S>.LSл#茳*"Jzw?0FI%J\hf,_BmęT% g&j=g Fjl2Qr-iP1S^j#]'N<5RDn3TMNXTaAkkčY.cUkF}k6~e]-$҄|pu&(Bb0|Pj')2G |Xf TJPa&R~ s+Ɉs\ ͽd8wxZ&`^d  I!s5)v\5DAwZr'rpґJsQ|c?Q'JᾮAЪu)jMMh=Sg?g5c2 w2TVh!TViMu۝u{nHX#Z#Ŀ{ow֑ݱu'.iKv.! IFQ58׉upm_QCPҽ3` vCpInwIu8JwI?K0;a7mHK[Yh4lVmoަk Vu޵^lYVc,1{X߂C ROvdhCŶsl[c4eR՘;[ssTw<_6bFR\jTr{ߞ]{뻾 =g}ᕃ^8%}uԜؒE# _@MmeflҹW΅ {޾x?һ͗w0iAp[2Ww4a4~Kcp)>Ţ?`0|ԚAo ʟdt3o\4_4<- 3Hz4)_ )+?ގ.~PmwQ wk üt(⢸}\cv9%WmMdnZ[rxŔAdRhd30H6#/ Zq*RyO!#E'RTRr8cj%:PcmFGmqv9n6hFL~G[ą Z?jчG#O"82xym|cI%XDBDT!MXEJAubX`~I>w0J_ Z_ ka~-0[O$oKBZ շ$E}K)TdBҧPPՒ*>^}_xӚ$Pг$4MQ&pLocEmԋ z)J=YZ0W>mji$4gTj1?yKO/6i ľB`Rϗ,S$.x${#'i}h [Z*Y?{2:~m>,~)r%)Q/ZM/gJtn9YU4֩֒k\E[1: ȰDe%z. Z٫zRm%SxgW% ]g1GWm-䉓" mT^b%R @\}/\A/i2ߑF8<3&G˻I&6DHDۑБ[{hT}\; idyT[Q4 Hp)'] + 3WJ{\ ޺Ei *UÕK^}W'q |EzYN|fqSX^V,.~H'qz2bU:"\r"aA话5vf&47!٭IҪCsG.n[8v y?WTpifpw֐5-sb8+n}pYV6A.qPA ,!l6YhrLKL@9r%ҥamھQcT*8zdt()JUu>~Z%(@M'x=f\wFU{N0{gV?_]̰93q}]+׺m1`q}Ln1`ŀ.uYXg Ո'Tgħ[ kɹʷ7OAs_u>3oF:W6yz;0JK'ŧK^mbEb6yM^lxJ^sIɋ/:y4!LE$\4HqWpDK3 JjZ]Oc(-bA_xɢ{7(Y lt+7۳qw}>{6nZz8b_|#y? n/n4us(cQR%;ibwfVWq\@;CC#M_[+Xui9:S;7C?@1 )}{].p@UjE+L#Ie/!L悄M4~t~`\D Ee3 P[QT] dm??VHtj U~jBRj|q0^jca.>&[2m ƶmA<{=i}ԛ(oTWL&qyI™HE)'B9|u"K!EACﱲF"VQ|m] Jo1p!qŨ[#q}9%6Ϋ76yuE؂$ʶCTSjW.¤z"0Ht[!PIrRDE3J[ p.R*U6aJp4ˍRV$qyeB<)19k( [C̥>gI 1LY,O$w&soºx$* -(un DeD@̥3EŤ" Pf`?osCb FnQl6D6*[ϷW{Z*u6f80?.\oV6X\=j!px~Oyv]m _6?toa5Of:Rl ^wm>A BnsR3X>283FpZTۤ/d! i8LiҿK-ҬPA^-jf~:_VV*I=+:xx2_iK7F*mQFE2 U]eL)TNQ'iCҋ6O?S7p4i`U\lv6sN gȊTӂf`?81T9Bk?(eiR8jUmO;$_e__BD {e'rx`b[~_fhtH/d.РW]ștJqVh\ !!D] [@qJw#!B16ĉ)$Xs! n ![N!DM!"!DpEkÀ|0Pwqq+2@+(ݰRe4ޔU&i}(T>P(lj"GҰTQd QdE  vQp 'Q3 DsPZmȄL3PLVC*%j1] X^Xh93jp`&)%y8ā6/n%Oi0,@ @yTM 0POÖ ;Dvba XIa@!ZH-xa@Hńu`@ X$g<|-||I>`Z8BX %> c}XJ#. t:S%;%#Tr3H 71ɰS9j{J"25=("6@UQ%P` حSy7wan8!9>ju["xwad aԟ5j6ܻ ivar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003225604415150373404017705 0ustar rootrootFeb 27 19:33:57 crc systemd[1]: Starting Kubernetes Kubelet... Feb 27 19:33:57 crc restorecon[4701]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:57 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 19:33:58 crc restorecon[4701]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 27 19:33:58 crc kubenswrapper[4839]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 19:33:58 crc kubenswrapper[4839]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 27 19:33:58 crc kubenswrapper[4839]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 19:33:58 crc kubenswrapper[4839]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 19:33:58 crc kubenswrapper[4839]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 27 19:33:58 crc kubenswrapper[4839]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.758169 4839 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763605 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763626 4839 feature_gate.go:330] unrecognized feature gate: Example Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763630 4839 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763635 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763639 4839 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763643 4839 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763647 4839 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763650 4839 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763654 4839 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763681 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763686 4839 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763690 4839 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763696 4839 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763702 4839 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763707 4839 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763711 4839 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763715 4839 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763720 4839 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763724 4839 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763727 4839 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763734 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763738 4839 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763742 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763745 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763749 4839 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763752 4839 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763756 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763759 4839 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763763 4839 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763767 4839 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763770 4839 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763774 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763777 4839 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763781 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763784 4839 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763787 4839 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763791 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763795 4839 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763799 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763802 4839 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763806 4839 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763809 4839 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763813 4839 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763817 4839 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763823 4839 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763829 4839 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763833 4839 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763837 4839 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763841 4839 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763845 4839 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763849 4839 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763853 4839 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763857 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763861 4839 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763865 4839 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763870 4839 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763873 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763877 4839 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763881 4839 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763885 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763888 4839 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763892 4839 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763895 4839 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763900 4839 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763910 4839 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763914 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763918 4839 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763921 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763925 4839 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763929 4839 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.763932 4839 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764753 4839 flags.go:64] FLAG: --address="0.0.0.0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764770 4839 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764779 4839 flags.go:64] FLAG: --anonymous-auth="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764786 4839 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764793 4839 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764799 4839 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764806 4839 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764813 4839 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764819 4839 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764824 4839 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764828 4839 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764835 4839 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764839 4839 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764844 4839 flags.go:64] FLAG: --cgroup-root="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764848 4839 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764852 4839 flags.go:64] FLAG: --client-ca-file="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764856 4839 flags.go:64] FLAG: --cloud-config="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764860 4839 flags.go:64] FLAG: --cloud-provider="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764864 4839 flags.go:64] FLAG: --cluster-dns="[]" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764869 4839 flags.go:64] FLAG: --cluster-domain="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764873 4839 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764877 4839 flags.go:64] FLAG: --config-dir="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764881 4839 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764885 4839 flags.go:64] FLAG: --container-log-max-files="5" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764890 4839 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764894 4839 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764898 4839 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764903 4839 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764907 4839 flags.go:64] FLAG: --contention-profiling="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764911 4839 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764915 4839 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764919 4839 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764925 4839 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764930 4839 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764934 4839 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764938 4839 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764942 4839 flags.go:64] FLAG: --enable-load-reader="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764946 4839 flags.go:64] FLAG: --enable-server="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764950 4839 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764955 4839 flags.go:64] FLAG: --event-burst="100" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764959 4839 flags.go:64] FLAG: --event-qps="50" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764963 4839 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764967 4839 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764971 4839 flags.go:64] FLAG: --eviction-hard="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764977 4839 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764982 4839 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764986 4839 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764992 4839 flags.go:64] FLAG: --eviction-soft="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.764997 4839 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765001 4839 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765005 4839 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765010 4839 flags.go:64] FLAG: --experimental-mounter-path="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765014 4839 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765017 4839 flags.go:64] FLAG: --fail-swap-on="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765062 4839 flags.go:64] FLAG: --feature-gates="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765068 4839 flags.go:64] FLAG: --file-check-frequency="20s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765072 4839 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765076 4839 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765081 4839 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765085 4839 flags.go:64] FLAG: --healthz-port="10248" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765090 4839 flags.go:64] FLAG: --help="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765095 4839 flags.go:64] FLAG: --hostname-override="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765099 4839 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765104 4839 flags.go:64] FLAG: --http-check-frequency="20s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765108 4839 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765113 4839 flags.go:64] FLAG: --image-credential-provider-config="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765117 4839 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765121 4839 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765125 4839 flags.go:64] FLAG: --image-service-endpoint="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765128 4839 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765133 4839 flags.go:64] FLAG: --kube-api-burst="100" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765137 4839 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765141 4839 flags.go:64] FLAG: --kube-api-qps="50" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765145 4839 flags.go:64] FLAG: --kube-reserved="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765149 4839 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765153 4839 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765157 4839 flags.go:64] FLAG: --kubelet-cgroups="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765162 4839 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765166 4839 flags.go:64] FLAG: --lock-file="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765169 4839 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765174 4839 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765178 4839 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765185 4839 flags.go:64] FLAG: --log-json-split-stream="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765190 4839 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765194 4839 flags.go:64] FLAG: --log-text-split-stream="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765198 4839 flags.go:64] FLAG: --logging-format="text" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765202 4839 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765206 4839 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765211 4839 flags.go:64] FLAG: --manifest-url="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765214 4839 flags.go:64] FLAG: --manifest-url-header="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765220 4839 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765224 4839 flags.go:64] FLAG: --max-open-files="1000000" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765229 4839 flags.go:64] FLAG: --max-pods="110" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765233 4839 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765238 4839 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765242 4839 flags.go:64] FLAG: --memory-manager-policy="None" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765246 4839 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765250 4839 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765254 4839 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765258 4839 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765268 4839 flags.go:64] FLAG: --node-status-max-images="50" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765272 4839 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765277 4839 flags.go:64] FLAG: --oom-score-adj="-999" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765281 4839 flags.go:64] FLAG: --pod-cidr="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765286 4839 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765292 4839 flags.go:64] FLAG: --pod-manifest-path="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765296 4839 flags.go:64] FLAG: --pod-max-pids="-1" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765300 4839 flags.go:64] FLAG: --pods-per-core="0" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765304 4839 flags.go:64] FLAG: --port="10250" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765309 4839 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765313 4839 flags.go:64] FLAG: --provider-id="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765317 4839 flags.go:64] FLAG: --qos-reserved="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765320 4839 flags.go:64] FLAG: --read-only-port="10255" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765325 4839 flags.go:64] FLAG: --register-node="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765329 4839 flags.go:64] FLAG: --register-schedulable="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765333 4839 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765340 4839 flags.go:64] FLAG: --registry-burst="10" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765345 4839 flags.go:64] FLAG: --registry-qps="5" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765350 4839 flags.go:64] FLAG: --reserved-cpus="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765355 4839 flags.go:64] FLAG: --reserved-memory="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765361 4839 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765365 4839 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765369 4839 flags.go:64] FLAG: --rotate-certificates="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765373 4839 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765378 4839 flags.go:64] FLAG: --runonce="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765382 4839 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765386 4839 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765391 4839 flags.go:64] FLAG: --seccomp-default="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765395 4839 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765400 4839 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765405 4839 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765409 4839 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765414 4839 flags.go:64] FLAG: --storage-driver-password="root" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765419 4839 flags.go:64] FLAG: --storage-driver-secure="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765423 4839 flags.go:64] FLAG: --storage-driver-table="stats" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765427 4839 flags.go:64] FLAG: --storage-driver-user="root" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765431 4839 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765435 4839 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765439 4839 flags.go:64] FLAG: --system-cgroups="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765443 4839 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765450 4839 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765454 4839 flags.go:64] FLAG: --tls-cert-file="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765458 4839 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765465 4839 flags.go:64] FLAG: --tls-min-version="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765469 4839 flags.go:64] FLAG: --tls-private-key-file="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765474 4839 flags.go:64] FLAG: --topology-manager-policy="none" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765480 4839 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765485 4839 flags.go:64] FLAG: --topology-manager-scope="container" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765490 4839 flags.go:64] FLAG: --v="2" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765501 4839 flags.go:64] FLAG: --version="false" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765507 4839 flags.go:64] FLAG: --vmodule="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765513 4839 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765518 4839 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765615 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765621 4839 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765627 4839 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765631 4839 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765635 4839 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765639 4839 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765643 4839 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765647 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765651 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765655 4839 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765703 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765709 4839 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765713 4839 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765717 4839 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765720 4839 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765724 4839 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765728 4839 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765731 4839 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765735 4839 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765739 4839 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765743 4839 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765746 4839 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765750 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765754 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765757 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765761 4839 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765765 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765768 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765772 4839 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765775 4839 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765778 4839 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765782 4839 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765785 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765789 4839 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765792 4839 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765796 4839 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765799 4839 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765803 4839 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765808 4839 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765811 4839 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765815 4839 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765819 4839 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765824 4839 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765829 4839 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765833 4839 feature_gate.go:330] unrecognized feature gate: Example Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765837 4839 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765840 4839 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765844 4839 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765847 4839 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765851 4839 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765855 4839 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765859 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765863 4839 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765866 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765870 4839 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765874 4839 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765878 4839 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765881 4839 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765885 4839 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765888 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765892 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765896 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765899 4839 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765903 4839 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765907 4839 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765910 4839 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765914 4839 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765918 4839 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765922 4839 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765932 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.765936 4839 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.765942 4839 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.773520 4839 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.773549 4839 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773608 4839 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773615 4839 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773619 4839 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773624 4839 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773629 4839 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773635 4839 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773640 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773644 4839 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773649 4839 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773652 4839 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773656 4839 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773704 4839 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773709 4839 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773714 4839 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773718 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773722 4839 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773726 4839 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773731 4839 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773735 4839 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773740 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773744 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773747 4839 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773751 4839 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773754 4839 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773758 4839 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773762 4839 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773765 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773769 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773772 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773776 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773779 4839 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773783 4839 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773786 4839 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773790 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773795 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773798 4839 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773803 4839 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773807 4839 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773811 4839 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773815 4839 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773819 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773822 4839 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773826 4839 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773830 4839 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773833 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773837 4839 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773840 4839 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773844 4839 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773847 4839 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773851 4839 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773855 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773859 4839 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773862 4839 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773865 4839 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773869 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773872 4839 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773876 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773879 4839 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773883 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773886 4839 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773890 4839 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773893 4839 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773896 4839 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773900 4839 feature_gate.go:330] unrecognized feature gate: Example Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773903 4839 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773907 4839 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773910 4839 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773914 4839 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773917 4839 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773920 4839 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.773924 4839 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.773931 4839 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774055 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774061 4839 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774065 4839 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774069 4839 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774073 4839 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774076 4839 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774080 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774085 4839 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774090 4839 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774095 4839 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774099 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774103 4839 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774107 4839 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774112 4839 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774115 4839 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774119 4839 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774123 4839 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774127 4839 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774130 4839 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774134 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774138 4839 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774141 4839 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774144 4839 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774148 4839 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774151 4839 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774155 4839 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774158 4839 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774162 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774165 4839 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774169 4839 feature_gate.go:330] unrecognized feature gate: Example Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774173 4839 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774176 4839 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774179 4839 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774183 4839 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774187 4839 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774191 4839 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774195 4839 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774199 4839 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774203 4839 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774207 4839 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774211 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774216 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774220 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774224 4839 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774227 4839 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774231 4839 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774234 4839 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774238 4839 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774241 4839 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774245 4839 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774248 4839 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774253 4839 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774258 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774261 4839 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774266 4839 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774271 4839 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774275 4839 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774278 4839 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774282 4839 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774285 4839 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774289 4839 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774292 4839 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774297 4839 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774302 4839 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774305 4839 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774309 4839 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774313 4839 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774316 4839 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774320 4839 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774323 4839 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.774327 4839 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.774333 4839 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.774471 4839 server.go:940] "Client rotation is on, will bootstrap in background" Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.777931 4839 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.780957 4839 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.781049 4839 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.782680 4839 server.go:997] "Starting client certificate rotation" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.782700 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.783242 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.810318 4839 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.810999 4839 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.815697 4839 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.835270 4839 log.go:25] "Validated CRI v1 runtime API" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.865277 4839 log.go:25] "Validated CRI v1 image API" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.867032 4839 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.870988 4839 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-27-19-29-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.871017 4839 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.885780 4839 manager.go:217] Machine: {Timestamp:2026-02-27 19:33:58.883564246 +0000 UTC m=+0.528434001 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:425800eb-b94c-4d28-9446-64bd59d9ebe7 BootID:da580aef-40a0-4e5a-b9b4-d01d8f39448d Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b5:c1:71 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b5:c1:71 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:8b:1a:2f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:68:08:23 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:85:13:f5 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ea:5e:c7 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:0a:20:47:bc:00:03 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9a:8a:b4:98:20:f3 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.886003 4839 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.886201 4839 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.887311 4839 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.887485 4839 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.887518 4839 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.887722 4839 topology_manager.go:138] "Creating topology manager with none policy" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.887732 4839 container_manager_linux.go:303] "Creating device plugin manager" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.888098 4839 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.888134 4839 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.888315 4839 state_mem.go:36] "Initialized new in-memory state store" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.888393 4839 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.891396 4839 kubelet.go:418] "Attempting to sync node with API server" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.891415 4839 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.891434 4839 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.891446 4839 kubelet.go:324] "Adding apiserver pod source" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.891456 4839 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.894625 4839 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.895476 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.895562 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.895700 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.895778 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.896052 4839 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.897972 4839 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899269 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899297 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899304 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899310 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899321 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899328 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899334 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899345 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899352 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899359 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899381 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.899387 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.901017 4839 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.901435 4839 server.go:1280] "Started kubelet" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.901579 4839 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.902763 4839 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.904302 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:58 crc systemd[1]: Started Kubernetes Kubelet. Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.906378 4839 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.912442 4839 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.912485 4839 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.912849 4839 server.go:460] "Adding debug handlers to kubelet server" Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.911479 4839 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.89:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189831769e436321 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.901408545 +0000 UTC m=+0.546278280,LastTimestamp:2026-02-27 19:33:58.901408545 +0000 UTC m=+0.546278280,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.914214 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.914286 4839 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.914271 4839 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.914309 4839 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.915178 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.915272 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.915941 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="200ms" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.918753 4839 factory.go:55] Registering systemd factory Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.918777 4839 factory.go:221] Registration of the systemd container factory successfully Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.919153 4839 factory.go:153] Registering CRI-O factory Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.919176 4839 factory.go:221] Registration of the crio container factory successfully Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.919254 4839 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.919288 4839 factory.go:103] Registering Raw factory Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.919307 4839 manager.go:1196] Started watching for new ooms in manager Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.920040 4839 manager.go:319] Starting recovery of all containers Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923045 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923090 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923104 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923116 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923127 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923139 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923151 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923162 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923176 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923190 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923229 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923242 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923253 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923268 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923280 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923290 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923301 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923337 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923349 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923363 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923376 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923388 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923400 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923411 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923424 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923438 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923473 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923489 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923503 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923516 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923528 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923541 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923554 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923567 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923577 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923589 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923604 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923617 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923629 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923640 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923728 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923745 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923757 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923769 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923781 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923793 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923803 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923816 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923827 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923837 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923850 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923862 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923877 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923889 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923902 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923913 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923924 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923938 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923949 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923961 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923973 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923985 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.923995 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924007 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924020 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924031 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924044 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924055 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924066 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924078 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924089 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924100 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924113 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924125 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924136 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924147 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924158 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924171 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924266 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924282 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924296 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924308 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924319 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924331 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924347 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924360 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924372 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924384 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924396 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924411 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924422 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924433 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924445 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924457 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924469 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.924482 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926531 4839 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926588 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926607 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926621 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926634 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926648 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926690 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926705 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926720 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926743 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926759 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926776 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926792 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926805 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926820 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926833 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926846 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926884 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926896 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926909 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926923 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926935 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926945 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926957 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.926969 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927022 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927070 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927089 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927103 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927117 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927130 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927144 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927159 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927171 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927184 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927203 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927216 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927231 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927245 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927257 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927270 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927547 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927577 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927597 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927608 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927619 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927636 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927647 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927675 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927686 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927699 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927713 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927725 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927741 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927756 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927767 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927782 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927792 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927804 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927819 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927830 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927843 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927855 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927865 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927879 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927889 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.927904 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929040 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929062 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929090 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929105 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929124 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929140 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929154 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929175 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929189 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929201 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929218 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929231 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929250 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929262 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929275 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929306 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929319 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929338 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929349 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929361 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929380 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929395 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929415 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929430 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929445 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929466 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929480 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929500 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929518 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929534 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929555 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929570 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929591 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929605 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929618 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929635 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929649 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929686 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929709 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929720 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929735 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929746 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929784 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929800 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929812 4839 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929822 4839 reconstruct.go:97] "Volume reconstruction finished" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.929830 4839 reconciler.go:26] "Reconciler: start to sync state" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.940405 4839 manager.go:324] Recovery completed Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.950775 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.952763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.952807 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.952839 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.953656 4839 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.953695 4839 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.953717 4839 state_mem.go:36] "Initialized new in-memory state store" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.961658 4839 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.964096 4839 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.964144 4839 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.964170 4839 kubelet.go:2335] "Starting kubelet main sync loop" Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.964221 4839 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 27 19:33:58 crc kubenswrapper[4839]: W0227 19:33:58.964743 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:58 crc kubenswrapper[4839]: E0227 19:33:58.964803 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.972130 4839 policy_none.go:49] "None policy: Start" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.972937 4839 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 27 19:33:58 crc kubenswrapper[4839]: I0227 19:33:58.972965 4839 state_mem.go:35] "Initializing new in-memory state store" Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.015308 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.021650 4839 manager.go:334] "Starting Device Plugin manager" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.021743 4839 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.021756 4839 server.go:79] "Starting device plugin registration server" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.022272 4839 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.022289 4839 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.022599 4839 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.022720 4839 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.022727 4839 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.030712 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.064747 4839 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.064858 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.066197 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.066266 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.066279 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.066506 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.066689 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.066758 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067648 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067702 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067715 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067871 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067912 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067935 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.067946 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068012 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068056 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068830 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068859 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068877 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068931 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069001 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069014 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069017 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069040 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.068950 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069930 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069981 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.069992 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070130 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070153 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070165 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070288 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070379 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070401 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.070984 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071007 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071019 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071145 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071173 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071551 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071578 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071590 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071762 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071780 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.071793 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.117440 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="400ms" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.122688 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.123731 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.123775 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.123791 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.123821 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.124404 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.89:6443: connect: connection refused" node="crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131533 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131560 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131576 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131593 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131610 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131691 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131707 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131771 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131787 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131800 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131813 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131851 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131879 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131917 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.131962 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233335 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233405 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233442 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233476 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233508 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233539 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233567 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233596 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233625 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233652 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233699 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233725 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233752 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233711 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233720 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233761 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233825 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233835 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233811 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233833 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233811 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233790 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233860 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233963 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.233992 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.234032 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.234061 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.234083 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.234112 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.234188 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.325356 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.326742 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.326774 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.326785 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.326807 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.327250 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.89:6443: connect: connection refused" node="crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.416456 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.437128 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.447570 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.461994 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a4b5c0e3bdfb015d4c07e75c611a0ad6d08ec209e8b7c8e84e203a819f899dcd WatchSource:0}: Error finding container a4b5c0e3bdfb015d4c07e75c611a0ad6d08ec209e8b7c8e84e203a819f899dcd: Status 404 returned error can't find the container with id a4b5c0e3bdfb015d4c07e75c611a0ad6d08ec209e8b7c8e84e203a819f899dcd Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.465450 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.471829 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.477427 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-66b3a702789c0fbd62cee182fdc64841698eb08edbca7e55050f671cd0692832 WatchSource:0}: Error finding container 66b3a702789c0fbd62cee182fdc64841698eb08edbca7e55050f671cd0692832: Status 404 returned error can't find the container with id 66b3a702789c0fbd62cee182fdc64841698eb08edbca7e55050f671cd0692832 Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.480473 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6e4f127b2468d01d0b243ce5638e853a13a30d46793e428098b91eb04beaf227 WatchSource:0}: Error finding container 6e4f127b2468d01d0b243ce5638e853a13a30d46793e428098b91eb04beaf227: Status 404 returned error can't find the container with id 6e4f127b2468d01d0b243ce5638e853a13a30d46793e428098b91eb04beaf227 Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.485150 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-8146a7f1d0e5bc277a22aaa05c5657aa083a2f7d0f9bc80e7e5dffd56aae5318 WatchSource:0}: Error finding container 8146a7f1d0e5bc277a22aaa05c5657aa083a2f7d0f9bc80e7e5dffd56aae5318: Status 404 returned error can't find the container with id 8146a7f1d0e5bc277a22aaa05c5657aa083a2f7d0f9bc80e7e5dffd56aae5318 Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.496850 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a163dc6d765a06aa9f8c34de9b83beb0173684a15c95fe7efbd2461de351636a WatchSource:0}: Error finding container a163dc6d765a06aa9f8c34de9b83beb0173684a15c95fe7efbd2461de351636a: Status 404 returned error can't find the container with id a163dc6d765a06aa9f8c34de9b83beb0173684a15c95fe7efbd2461de351636a Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.518494 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="800ms" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.728255 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.729718 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.729766 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.729778 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.729806 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.730228 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.89:6443: connect: connection refused" node="crc" Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.823806 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.823877 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.905328 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:59 crc kubenswrapper[4839]: W0227 19:33:59.961835 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:33:59 crc kubenswrapper[4839]: E0227 19:33:59.961892 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.969956 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8146a7f1d0e5bc277a22aaa05c5657aa083a2f7d0f9bc80e7e5dffd56aae5318"} Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.971174 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6e4f127b2468d01d0b243ce5638e853a13a30d46793e428098b91eb04beaf227"} Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.971884 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"66b3a702789c0fbd62cee182fdc64841698eb08edbca7e55050f671cd0692832"} Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.972605 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a4b5c0e3bdfb015d4c07e75c611a0ad6d08ec209e8b7c8e84e203a819f899dcd"} Feb 27 19:33:59 crc kubenswrapper[4839]: I0227 19:33:59.973345 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a163dc6d765a06aa9f8c34de9b83beb0173684a15c95fe7efbd2461de351636a"} Feb 27 19:34:00 crc kubenswrapper[4839]: W0227 19:34:00.215762 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:34:00 crc kubenswrapper[4839]: E0227 19:34:00.215846 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:34:00 crc kubenswrapper[4839]: E0227 19:34:00.319165 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="1.6s" Feb 27 19:34:00 crc kubenswrapper[4839]: W0227 19:34:00.439291 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:34:00 crc kubenswrapper[4839]: E0227 19:34:00.439373 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.531329 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.532795 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.532853 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.532873 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.532912 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:00 crc kubenswrapper[4839]: E0227 19:34:00.533619 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.89:6443: connect: connection refused" node="crc" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.865285 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 19:34:00 crc kubenswrapper[4839]: E0227 19:34:00.866488 4839 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.905454 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.979823 4839 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8ebeca656e1b52ae5de19c5e2f8fe7adab3294d03c66bf4c30e7d3d7e2070c55" exitCode=0 Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.979898 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8ebeca656e1b52ae5de19c5e2f8fe7adab3294d03c66bf4c30e7d3d7e2070c55"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.980025 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.982618 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.982739 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.982770 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.987138 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.987162 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.987211 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.987231 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.987248 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.988144 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.988203 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.988220 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.990282 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2" exitCode=0 Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.990328 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.990430 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.990892 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.990915 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.990922 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.992439 4839 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba" exitCode=0 Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.992530 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.992800 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.992972 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.993194 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.993213 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.993224 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.993942 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.993961 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.993969 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.996330 4839 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42" exitCode=0 Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.996359 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42"} Feb 27 19:34:00 crc kubenswrapper[4839]: I0227 19:34:00.996519 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:01 crc kubenswrapper[4839]: I0227 19:34:00.999841 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:01 crc kubenswrapper[4839]: I0227 19:34:00.999856 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:01 crc kubenswrapper[4839]: I0227 19:34:00.999864 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:01 crc kubenswrapper[4839]: I0227 19:34:01.905082 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:34:01 crc kubenswrapper[4839]: E0227 19:34:01.920203 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="3.2s" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.000888 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cb207eb4a6d2a0dfe3cc465f4a9932bb272534cd99be6d5d31a887b72dfcc062"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.000956 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"64633ef809e1b51516a270afe403fcfc5134006ccc054f436d23cfc38071d5bb"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.000977 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f7d8881df9fc1481b72c4c1cfc7175b4685e0346664391961b45ca1135fab6a1"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.000913 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.002206 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.002242 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.002254 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.005556 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.005612 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.005632 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.005651 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.008350 4839 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6" exitCode=0 Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.008422 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.008616 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.009474 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.009510 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.009524 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.011370 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.011886 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.012227 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130"} Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.012880 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.012946 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.012960 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.013031 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.013051 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.013062 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:02 crc kubenswrapper[4839]: W0227 19:34:02.115935 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.89:6443: connect: connection refused Feb 27 19:34:02 crc kubenswrapper[4839]: E0227 19:34:02.116023 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.89:6443: connect: connection refused" logger="UnhandledError" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.134607 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.136210 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.136243 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.136252 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:02 crc kubenswrapper[4839]: I0227 19:34:02.136270 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:02 crc kubenswrapper[4839]: E0227 19:34:02.136643 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.89:6443: connect: connection refused" node="crc" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.018961 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e6ed37b1bcc9572fd436931e7bd7712987cd81af6c6654868d681dd0d780bfe"} Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.019410 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.020766 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.020854 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.020880 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.024172 4839 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105" exitCode=0 Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.024245 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105"} Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.024293 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.024398 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.024302 4839 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.024512 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.026262 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.026326 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.026371 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.027107 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.027153 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.027189 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.030189 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.030247 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:03 crc kubenswrapper[4839]: I0227 19:34:03.030271 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.030857 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff"} Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.030907 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f"} Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.030932 4839 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.031000 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.030926 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799"} Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.031064 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec"} Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.032344 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.032395 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.032407 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.939701 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.939930 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.941375 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.941440 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:04 crc kubenswrapper[4839]: I0227 19:34:04.941479 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.040208 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce"} Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.040454 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.042924 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.042999 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.043018 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.209025 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.209248 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.211062 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.211141 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.211161 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.220001 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.220175 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.221636 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.221716 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.221735 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.250340 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.337771 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.339764 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.339830 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.339850 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.339888 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.567242 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.873877 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.983620 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.983894 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.985894 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.985967 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:05 crc kubenswrapper[4839]: I0227 19:34:05.985993 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.043384 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.043408 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.044767 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.044830 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.044853 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.045193 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.045218 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:06 crc kubenswrapper[4839]: I0227 19:34:06.045229 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:07 crc kubenswrapper[4839]: I0227 19:34:07.046230 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:07 crc kubenswrapper[4839]: I0227 19:34:07.050608 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:07 crc kubenswrapper[4839]: I0227 19:34:07.050709 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:07 crc kubenswrapper[4839]: I0227 19:34:07.050746 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:07 crc kubenswrapper[4839]: I0227 19:34:07.940362 4839 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:34:07 crc kubenswrapper[4839]: I0227 19:34:07.940844 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.240654 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.240932 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.242576 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.242646 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.242709 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.248412 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.596475 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.597057 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.598782 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.598835 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:08 crc kubenswrapper[4839]: I0227 19:34:08.598854 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:09 crc kubenswrapper[4839]: E0227 19:34:09.031071 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.051843 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.051931 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.053265 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.053306 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.053332 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.189251 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.189501 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.191020 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.191095 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:09 crc kubenswrapper[4839]: I0227 19:34:09.191111 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:10 crc kubenswrapper[4839]: I0227 19:34:10.055988 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:10 crc kubenswrapper[4839]: I0227 19:34:10.057579 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:10 crc kubenswrapper[4839]: I0227 19:34:10.057660 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:10 crc kubenswrapper[4839]: I0227 19:34:10.057720 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:10 crc kubenswrapper[4839]: I0227 19:34:10.063167 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:11 crc kubenswrapper[4839]: I0227 19:34:11.058365 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:11 crc kubenswrapper[4839]: I0227 19:34:11.059396 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:11 crc kubenswrapper[4839]: I0227 19:34:11.059449 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:11 crc kubenswrapper[4839]: I0227 19:34:11.059467 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:12 crc kubenswrapper[4839]: W0227 19:34:12.624832 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 19:34:12 crc kubenswrapper[4839]: I0227 19:34:12.624935 4839 trace.go:236] Trace[1324500036]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 19:34:02.623) (total time: 10001ms): Feb 27 19:34:12 crc kubenswrapper[4839]: Trace[1324500036]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (19:34:12.624) Feb 27 19:34:12 crc kubenswrapper[4839]: Trace[1324500036]: [10.001578517s] [10.001578517s] END Feb 27 19:34:12 crc kubenswrapper[4839]: E0227 19:34:12.624963 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 19:34:12 crc kubenswrapper[4839]: W0227 19:34:12.678876 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 19:34:12 crc kubenswrapper[4839]: I0227 19:34:12.679027 4839 trace.go:236] Trace[504627309]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 19:34:02.677) (total time: 10001ms): Feb 27 19:34:12 crc kubenswrapper[4839]: Trace[504627309]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (19:34:12.678) Feb 27 19:34:12 crc kubenswrapper[4839]: Trace[504627309]: [10.001812892s] [10.001812892s] END Feb 27 19:34:12 crc kubenswrapper[4839]: E0227 19:34:12.679063 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 19:34:12 crc kubenswrapper[4839]: W0227 19:34:12.779937 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 19:34:12 crc kubenswrapper[4839]: I0227 19:34:12.780032 4839 trace.go:236] Trace[1451081211]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 19:34:02.778) (total time: 10001ms): Feb 27 19:34:12 crc kubenswrapper[4839]: Trace[1451081211]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (19:34:12.779) Feb 27 19:34:12 crc kubenswrapper[4839]: Trace[1451081211]: [10.00167605s] [10.00167605s] END Feb 27 19:34:12 crc kubenswrapper[4839]: E0227 19:34:12.780057 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 19:34:12 crc kubenswrapper[4839]: I0227 19:34:12.905833 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 27 19:34:13 crc kubenswrapper[4839]: E0227 19:34:13.097888 4839 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189831769e436321 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.901408545 +0000 UTC m=+0.546278280,LastTimestamp:2026-02-27 19:33:58.901408545 +0000 UTC m=+0.546278280,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:13 crc kubenswrapper[4839]: W0227 19:34:13.102135 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z Feb 27 19:34:13 crc kubenswrapper[4839]: E0227 19:34:13.102199 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 19:34:13 crc kubenswrapper[4839]: E0227 19:34:13.103712 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 19:34:13 crc kubenswrapper[4839]: E0227 19:34:13.107549 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 27 19:34:13 crc kubenswrapper[4839]: I0227 19:34:13.110357 4839 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 19:34:13 crc kubenswrapper[4839]: I0227 19:34:13.110433 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 19:34:13 crc kubenswrapper[4839]: E0227 19:34:13.112854 4839 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 19:34:13 crc kubenswrapper[4839]: I0227 19:34:13.116270 4839 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 19:34:13 crc kubenswrapper[4839]: I0227 19:34:13.116323 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 19:34:13 crc kubenswrapper[4839]: I0227 19:34:13.909944 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:13Z is after 2026-02-23T05:33:13Z Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.068436 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.071316 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8e6ed37b1bcc9572fd436931e7bd7712987cd81af6c6654868d681dd0d780bfe" exitCode=255 Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.071369 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8e6ed37b1bcc9572fd436931e7bd7712987cd81af6c6654868d681dd0d780bfe"} Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.071569 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.072818 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.072947 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.073035 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.073728 4839 scope.go:117] "RemoveContainer" containerID="8e6ed37b1bcc9572fd436931e7bd7712987cd81af6c6654868d681dd0d780bfe" Feb 27 19:34:14 crc kubenswrapper[4839]: I0227 19:34:14.912889 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:14Z is after 2026-02-23T05:33:13Z Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.077563 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.078361 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.080888 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" exitCode=255 Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.080947 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc"} Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.081008 4839 scope.go:117] "RemoveContainer" containerID="8e6ed37b1bcc9572fd436931e7bd7712987cd81af6c6654868d681dd0d780bfe" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.081116 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.081912 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.081956 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.081973 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.082862 4839 scope.go:117] "RemoveContainer" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" Feb 27 19:34:15 crc kubenswrapper[4839]: E0227 19:34:15.083136 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.220109 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.576824 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.880772 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:15 crc kubenswrapper[4839]: I0227 19:34:15.910297 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:15Z is after 2026-02-23T05:33:13Z Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.085692 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.088646 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.089356 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.089389 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.089398 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.089857 4839 scope.go:117] "RemoveContainer" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" Feb 27 19:34:16 crc kubenswrapper[4839]: E0227 19:34:16.090005 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.910550 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:16Z is after 2026-02-23T05:33:13Z Feb 27 19:34:16 crc kubenswrapper[4839]: I0227 19:34:16.947180 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:17 crc kubenswrapper[4839]: W0227 19:34:17.029263 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z Feb 27 19:34:17 crc kubenswrapper[4839]: E0227 19:34:17.029346 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.091997 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.093249 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.093323 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.093352 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.094327 4839 scope.go:117] "RemoveContainer" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" Feb 27 19:34:17 crc kubenswrapper[4839]: E0227 19:34:17.094712 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:17 crc kubenswrapper[4839]: W0227 19:34:17.419122 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z Feb 27 19:34:17 crc kubenswrapper[4839]: E0227 19:34:17.419249 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 19:34:17 crc kubenswrapper[4839]: W0227 19:34:17.563855 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z Feb 27 19:34:17 crc kubenswrapper[4839]: E0227 19:34:17.563964 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.910648 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:17Z is after 2026-02-23T05:33:13Z Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.941441 4839 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:34:17 crc kubenswrapper[4839]: I0227 19:34:17.941638 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:34:18 crc kubenswrapper[4839]: I0227 19:34:18.094226 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:18 crc kubenswrapper[4839]: I0227 19:34:18.095533 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:18 crc kubenswrapper[4839]: I0227 19:34:18.095843 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:18 crc kubenswrapper[4839]: I0227 19:34:18.096025 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:18 crc kubenswrapper[4839]: I0227 19:34:18.096970 4839 scope.go:117] "RemoveContainer" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" Feb 27 19:34:18 crc kubenswrapper[4839]: E0227 19:34:18.097389 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:18 crc kubenswrapper[4839]: I0227 19:34:18.909119 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:34:18Z is after 2026-02-23T05:33:13Z Feb 27 19:34:19 crc kubenswrapper[4839]: E0227 19:34:19.031261 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.229981 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.230274 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.232009 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.232073 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.232097 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.244951 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.504176 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.506307 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.506395 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.506421 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.506471 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:19 crc kubenswrapper[4839]: E0227 19:34:19.513609 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:34:19 crc kubenswrapper[4839]: E0227 19:34:19.513950 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:34:19 crc kubenswrapper[4839]: I0227 19:34:19.908884 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:20 crc kubenswrapper[4839]: I0227 19:34:20.098593 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:20 crc kubenswrapper[4839]: I0227 19:34:20.104466 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:20 crc kubenswrapper[4839]: I0227 19:34:20.104962 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:20 crc kubenswrapper[4839]: I0227 19:34:20.105177 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:20 crc kubenswrapper[4839]: I0227 19:34:20.912448 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:21 crc kubenswrapper[4839]: I0227 19:34:21.139326 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 19:34:21 crc kubenswrapper[4839]: I0227 19:34:21.159460 4839 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 19:34:21 crc kubenswrapper[4839]: I0227 19:34:21.933518 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:22 crc kubenswrapper[4839]: I0227 19:34:22.910716 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.105066 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189831769e436321 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.901408545 +0000 UTC m=+0.546278280,LastTimestamp:2026-02-27 19:33:58.901408545 +0000 UTC m=+0.546278280,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.110130 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.111425 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.115468 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.119759 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a5a5de65 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:59.025303141 +0000 UTC m=+0.670172866,LastTimestamp:2026-02-27 19:33:59.025303141 +0000 UTC m=+0.670172866,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.124534 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.066249924 +0000 UTC m=+0.711119669,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.129010 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.066274953 +0000 UTC m=+0.711144698,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.135363 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a154472a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:59.066286102 +0000 UTC m=+0.711155847,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.140205 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.067696438 +0000 UTC m=+0.712566183,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.144118 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.067710847 +0000 UTC m=+0.712580592,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.150026 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a154472a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:59.067721416 +0000 UTC m=+0.712591171,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.156872 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.067929374 +0000 UTC m=+0.712799129,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.163405 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.067942473 +0000 UTC m=+0.712812218,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.168592 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a154472a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:59.067953263 +0000 UTC m=+0.712823018,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.175501 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.068847289 +0000 UTC m=+0.713717024,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.182320 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.068867058 +0000 UTC m=+0.713736793,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.186696 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a154472a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:59.068884317 +0000 UTC m=+0.713754052,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.190800 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.06899494 +0000 UTC m=+0.713864685,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.196691 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.069010109 +0000 UTC m=+0.713879854,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.200540 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a154472a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:59.069020619 +0000 UTC m=+0.713890364,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.204030 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.069974732 +0000 UTC m=+0.714844477,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.209791 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.069989011 +0000 UTC m=+0.714858766,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.218410 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a154472a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a154472a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952847146 +0000 UTC m=+0.597716881,LastTimestamp:2026-02-27 19:33:59.06999892 +0000 UTC m=+0.714868665,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.224288 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153706c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153706c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952792172 +0000 UTC m=+0.597661917,LastTimestamp:2026-02-27 19:33:59.070146542 +0000 UTC m=+0.715016287,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.227954 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18983176a153c9a6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18983176a153c9a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:58.952815014 +0000 UTC m=+0.597684759,LastTimestamp:2026-02-27 19:33:59.070159631 +0000 UTC m=+0.715029376,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.236152 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18983176c02340b5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:59.469727925 +0000 UTC m=+1.114597680,LastTimestamp:2026-02-27 19:33:59.469727925 +0000 UTC m=+1.114597680,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.240900 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983176c0ba3abd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:59.479622333 +0000 UTC m=+1.124492068,LastTimestamp:2026-02-27 19:33:59.479622333 +0000 UTC m=+1.124492068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.246812 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983176c0fc421b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:59.483949595 +0000 UTC m=+1.128819340,LastTimestamp:2026-02-27 19:33:59.483949595 +0000 UTC m=+1.128819340,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.250275 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176c155078b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:59.489767307 +0000 UTC m=+1.134637052,LastTimestamp:2026-02-27 19:33:59.489767307 +0000 UTC m=+1.134637052,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.254473 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18983176c22fdbac openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:33:59.50410846 +0000 UTC m=+1.148978215,LastTimestamp:2026-02-27 19:33:59.50410846 +0000 UTC m=+1.148978215,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.260505 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18983176dfec71be openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.00300691 +0000 UTC m=+1.647876645,LastTimestamp:2026-02-27 19:34:00.00300691 +0000 UTC m=+1.647876645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.266922 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176dff1f7f4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.003368948 +0000 UTC m=+1.648238693,LastTimestamp:2026-02-27 19:34:00.003368948 +0000 UTC m=+1.648238693,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.273399 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983176dff313e2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.003441634 +0000 UTC m=+1.648311359,LastTimestamp:2026-02-27 19:34:00.003441634 +0000 UTC m=+1.648311359,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.279622 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18983176dff4ccad openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.003554477 +0000 UTC m=+1.648424212,LastTimestamp:2026-02-27 19:34:00.003554477 +0000 UTC m=+1.648424212,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.284451 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983176dff78259 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.003732057 +0000 UTC m=+1.648601802,LastTimestamp:2026-02-27 19:34:00.003732057 +0000 UTC m=+1.648601802,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.290521 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18983176e0a5b007 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.015147015 +0000 UTC m=+1.660016760,LastTimestamp:2026-02-27 19:34:00.015147015 +0000 UTC m=+1.660016760,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.294806 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176e0af7200 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.015786496 +0000 UTC m=+1.660656231,LastTimestamp:2026-02-27 19:34:00.015786496 +0000 UTC m=+1.660656231,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.299351 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18983176e0b9d33a openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.016466746 +0000 UTC m=+1.661336491,LastTimestamp:2026-02-27 19:34:00.016466746 +0000 UTC m=+1.661336491,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.303570 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176e0bf2401 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.016815105 +0000 UTC m=+1.661684850,LastTimestamp:2026-02-27 19:34:00.016815105 +0000 UTC m=+1.661684850,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.306982 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983176e0f7e023 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.020533283 +0000 UTC m=+1.665403028,LastTimestamp:2026-02-27 19:34:00.020533283 +0000 UTC m=+1.665403028,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.313530 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983176e11fe594 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.023156116 +0000 UTC m=+1.668025861,LastTimestamp:2026-02-27 19:34:00.023156116 +0000 UTC m=+1.668025861,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.317267 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176f0a44901 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.283490561 +0000 UTC m=+1.928360296,LastTimestamp:2026-02-27 19:34:00.283490561 +0000 UTC m=+1.928360296,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.320262 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176f1720705 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.296974085 +0000 UTC m=+1.941843830,LastTimestamp:2026-02-27 19:34:00.296974085 +0000 UTC m=+1.941843830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.322511 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176f184cb7f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.298204031 +0000 UTC m=+1.943073786,LastTimestamp:2026-02-27 19:34:00.298204031 +0000 UTC m=+1.943073786,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.327392 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176ff05cfa4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.524763044 +0000 UTC m=+2.169632819,LastTimestamp:2026-02-27 19:34:00.524763044 +0000 UTC m=+2.169632819,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.329467 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176ffacc152 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.53570389 +0000 UTC m=+2.180573645,LastTimestamp:2026-02-27 19:34:00.53570389 +0000 UTC m=+2.180573645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.336157 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176ffbaec19 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.536632345 +0000 UTC m=+2.181502100,LastTimestamp:2026-02-27 19:34:00.536632345 +0000 UTC m=+2.181502100,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.342219 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189831770b97afe3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.735649763 +0000 UTC m=+2.380519498,LastTimestamp:2026-02-27 19:34:00.735649763 +0000 UTC m=+2.380519498,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.348483 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189831770c1f5c62 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.744541282 +0000 UTC m=+2.389411017,LastTimestamp:2026-02-27 19:34:00.744541282 +0000 UTC m=+2.389411017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.356333 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189831771a6c1a35 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.984451637 +0000 UTC m=+2.629321412,LastTimestamp:2026-02-27 19:34:00.984451637 +0000 UTC m=+2.629321412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.361608 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831771aea751a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.992732442 +0000 UTC m=+2.637602207,LastTimestamp:2026-02-27 19:34:00.992732442 +0000 UTC m=+2.637602207,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.368051 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189831771b0b0688 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.994866824 +0000 UTC m=+2.639736599,LastTimestamp:2026-02-27 19:34:00.994866824 +0000 UTC m=+2.639736599,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.374847 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189831771b614cb6 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.000520886 +0000 UTC m=+2.645390631,LastTimestamp:2026-02-27 19:34:01.000520886 +0000 UTC m=+2.645390631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.379403 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18983177287be8d3 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.220368595 +0000 UTC m=+2.865238320,LastTimestamp:2026-02-27 19:34:01.220368595 +0000 UTC m=+2.865238320,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.384828 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898317728ac262d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.223530029 +0000 UTC m=+2.868399754,LastTimestamp:2026-02-27 19:34:01.223530029 +0000 UTC m=+2.868399754,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.389700 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898317728b83728 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.224320808 +0000 UTC m=+2.869190543,LastTimestamp:2026-02-27 19:34:01.224320808 +0000 UTC m=+2.869190543,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.394324 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898317728dc6a0d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.226693133 +0000 UTC m=+2.871562868,LastTimestamp:2026-02-27 19:34:01.226693133 +0000 UTC m=+2.871562868,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.401257 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898317729270d2e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.231584558 +0000 UTC m=+2.876454333,LastTimestamp:2026-02-27 19:34:01.231584558 +0000 UTC m=+2.876454333,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.407759 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898317729414997 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.233303959 +0000 UTC m=+2.878173694,LastTimestamp:2026-02-27 19:34:01.233303959 +0000 UTC m=+2.878173694,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.414098 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983177296ee663 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.236293219 +0000 UTC m=+2.881162954,LastTimestamp:2026-02-27 19:34:01.236293219 +0000 UTC m=+2.881162954,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.420846 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983177297bb298 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.237131928 +0000 UTC m=+2.882001683,LastTimestamp:2026-02-27 19:34:01.237131928 +0000 UTC m=+2.882001683,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.427069 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898317729d78727 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.243150119 +0000 UTC m=+2.888019865,LastTimestamp:2026-02-27 19:34:01.243150119 +0000 UTC m=+2.888019865,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.433135 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189831772a2ed680 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.248872064 +0000 UTC m=+2.893741819,LastTimestamp:2026-02-27 19:34:01.248872064 +0000 UTC m=+2.893741819,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.438920 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898317734670e27 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.420328487 +0000 UTC m=+3.065198242,LastTimestamp:2026-02-27 19:34:01.420328487 +0000 UTC m=+3.065198242,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.443553 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898317735738be0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.43792432 +0000 UTC m=+3.082794055,LastTimestamp:2026-02-27 19:34:01.43792432 +0000 UTC m=+3.082794055,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.447635 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189831773581fcae openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.438870702 +0000 UTC m=+3.083740437,LastTimestamp:2026-02-27 19:34:01.438870702 +0000 UTC m=+3.083740437,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.451205 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983177385de4ac openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.486836908 +0000 UTC m=+3.131706653,LastTimestamp:2026-02-27 19:34:01.486836908 +0000 UTC m=+3.131706653,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.455771 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983177393cb1fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.501438461 +0000 UTC m=+3.146308196,LastTimestamp:2026-02-27 19:34:01.501438461 +0000 UTC m=+3.146308196,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.459591 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983177394c2b5e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.502452574 +0000 UTC m=+3.147322349,LastTimestamp:2026-02-27 19:34:01.502452574 +0000 UTC m=+3.147322349,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.464873 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189831773f4afa9f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.603037855 +0000 UTC m=+3.247907600,LastTimestamp:2026-02-27 19:34:01.603037855 +0000 UTC m=+3.247907600,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.469207 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189831773fcdeffa openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.611620346 +0000 UTC m=+3.256490081,LastTimestamp:2026-02-27 19:34:01.611620346 +0000 UTC m=+3.256490081,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.474003 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983177421c7298 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.650320024 +0000 UTC m=+3.295189749,LastTimestamp:2026-02-27 19:34:01.650320024 +0000 UTC m=+3.295189749,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.479818 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898317742bf2c9b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.660984475 +0000 UTC m=+3.305854210,LastTimestamp:2026-02-27 19:34:01.660984475 +0000 UTC m=+3.305854210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.484156 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898317742cd3f62 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.661906786 +0000 UTC m=+3.306776521,LastTimestamp:2026-02-27 19:34:01.661906786 +0000 UTC m=+3.306776521,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.489705 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831774c576c7a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.821957242 +0000 UTC m=+3.466826997,LastTimestamp:2026-02-27 19:34:01.821957242 +0000 UTC m=+3.466826997,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.493841 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831774cd6fdec openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.830317548 +0000 UTC m=+3.475187313,LastTimestamp:2026-02-27 19:34:01.830317548 +0000 UTC m=+3.475187313,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.498884 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831774ce72023 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.831374883 +0000 UTC m=+3.476244658,LastTimestamp:2026-02-27 19:34:01.831374883 +0000 UTC m=+3.476244658,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.505621 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898317757b0cdf8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.01236428 +0000 UTC m=+3.657234025,LastTimestamp:2026-02-27 19:34:02.01236428 +0000 UTC m=+3.657234025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.511161 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831775ac2e6b5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.063881909 +0000 UTC m=+3.708751704,LastTimestamp:2026-02-27 19:34:02.063881909 +0000 UTC m=+3.708751704,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.519568 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831775bd6661b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.081936923 +0000 UTC m=+3.726806668,LastTimestamp:2026-02-27 19:34:02.081936923 +0000 UTC m=+3.726806668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.525379 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898317762fb090d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.201778445 +0000 UTC m=+3.846648170,LastTimestamp:2026-02-27 19:34:02.201778445 +0000 UTC m=+3.846648170,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.529421 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898317763a073dd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.212619229 +0000 UTC m=+3.857488954,LastTimestamp:2026-02-27 19:34:02.212619229 +0000 UTC m=+3.857488954,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.534736 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898317794652c89 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.030817929 +0000 UTC m=+4.675687674,LastTimestamp:2026-02-27 19:34:03.030817929 +0000 UTC m=+4.675687674,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.540366 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177a34ccdf9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.280879097 +0000 UTC m=+4.925748872,LastTimestamp:2026-02-27 19:34:03.280879097 +0000 UTC m=+4.925748872,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.546075 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177a3f7417d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.292049789 +0000 UTC m=+4.936919554,LastTimestamp:2026-02-27 19:34:03.292049789 +0000 UTC m=+4.936919554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.550778 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177a408f820 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.293210656 +0000 UTC m=+4.938080421,LastTimestamp:2026-02-27 19:34:03.293210656 +0000 UTC m=+4.938080421,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.555383 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177af6745ce openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.483940302 +0000 UTC m=+5.128810047,LastTimestamp:2026-02-27 19:34:03.483940302 +0000 UTC m=+5.128810047,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.559799 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177b02df1f1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.496960497 +0000 UTC m=+5.141830232,LastTimestamp:2026-02-27 19:34:03.496960497 +0000 UTC m=+5.141830232,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.565814 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177b042235c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.498283868 +0000 UTC m=+5.143153613,LastTimestamp:2026-02-27 19:34:03.498283868 +0000 UTC m=+5.143153613,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.573603 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177bc486480 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.700020352 +0000 UTC m=+5.344890097,LastTimestamp:2026-02-27 19:34:03.700020352 +0000 UTC m=+5.344890097,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.580308 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177bd28a6c1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.714717377 +0000 UTC m=+5.359587122,LastTimestamp:2026-02-27 19:34:03.714717377 +0000 UTC m=+5.359587122,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.587164 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177bd3ea141 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.716157761 +0000 UTC m=+5.361027506,LastTimestamp:2026-02-27 19:34:03.716157761 +0000 UTC m=+5.361027506,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.594701 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177cc7c170b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.971843851 +0000 UTC m=+5.616713596,LastTimestamp:2026-02-27 19:34:03.971843851 +0000 UTC m=+5.616713596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.601070 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177cd85bcbe openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.98925331 +0000 UTC m=+5.634123055,LastTimestamp:2026-02-27 19:34:03.98925331 +0000 UTC m=+5.634123055,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.606332 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177cd955f74 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:03.990278004 +0000 UTC m=+5.635147749,LastTimestamp:2026-02-27 19:34:03.990278004 +0000 UTC m=+5.635147749,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.611792 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177db4324cb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:04.219770059 +0000 UTC m=+5.864639804,LastTimestamp:2026-02-27 19:34:04.219770059 +0000 UTC m=+5.864639804,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.617705 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18983177dc474ff5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:04.236820469 +0000 UTC m=+5.881690224,LastTimestamp:2026-02-27 19:34:04.236820469 +0000 UTC m=+5.881690224,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.626305 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 19:34:23 crc kubenswrapper[4839]: &Event{ObjectMeta:{kube-controller-manager-crc.18983178b90d9004 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Feb 27 19:34:23 crc kubenswrapper[4839]: body: Feb 27 19:34:23 crc kubenswrapper[4839]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:07.940800516 +0000 UTC m=+9.585670321,LastTimestamp:2026-02-27 19:34:07.940800516 +0000 UTC m=+9.585670321,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 19:34:23 crc kubenswrapper[4839]: > Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.629556 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983178b911767a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:07.941056122 +0000 UTC m=+9.585925937,LastTimestamp:2026-02-27 19:34:07.941056122 +0000 UTC m=+9.585925937,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.634609 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 19:34:23 crc kubenswrapper[4839]: &Event{ObjectMeta:{kube-apiserver-crc.18983179ed2f8dd3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 19:34:23 crc kubenswrapper[4839]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 19:34:23 crc kubenswrapper[4839]: Feb 27 19:34:23 crc kubenswrapper[4839]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:13.110410707 +0000 UTC m=+14.755280442,LastTimestamp:2026-02-27 19:34:13.110410707 +0000 UTC m=+14.755280442,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 19:34:23 crc kubenswrapper[4839]: > Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.641518 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983179ed304f94 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:13.110460308 +0000 UTC m=+14.755330043,LastTimestamp:2026-02-27 19:34:13.110460308 +0000 UTC m=+14.755330043,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.647908 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18983179ed2f8dd3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 19:34:23 crc kubenswrapper[4839]: &Event{ObjectMeta:{kube-apiserver-crc.18983179ed2f8dd3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 19:34:23 crc kubenswrapper[4839]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 19:34:23 crc kubenswrapper[4839]: Feb 27 19:34:23 crc kubenswrapper[4839]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:13.110410707 +0000 UTC m=+14.755280442,LastTimestamp:2026-02-27 19:34:13.116307765 +0000 UTC m=+14.761177490,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 19:34:23 crc kubenswrapper[4839]: > Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.654568 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18983179ed304f94\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18983179ed304f94 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:13.110460308 +0000 UTC m=+14.755330043,LastTimestamp:2026-02-27 19:34:13.116341646 +0000 UTC m=+14.761211381,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.664156 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189831774ce72023\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831774ce72023 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:01.831374883 +0000 UTC m=+3.476244658,LastTimestamp:2026-02-27 19:34:14.074914929 +0000 UTC m=+15.719784704,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.670151 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189831775ac2e6b5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831775ac2e6b5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.063881909 +0000 UTC m=+3.708751704,LastTimestamp:2026-02-27 19:34:14.261830936 +0000 UTC m=+15.906700661,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.676177 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189831775bd6661b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189831775bd6661b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:02.081936923 +0000 UTC m=+3.726806668,LastTimestamp:2026-02-27 19:34:14.269585138 +0000 UTC m=+15.914454873,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.681071 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 19:34:23 crc kubenswrapper[4839]: &Event{ObjectMeta:{kube-controller-manager-crc.1898317b0d25de2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 19:34:23 crc kubenswrapper[4839]: body: Feb 27 19:34:23 crc kubenswrapper[4839]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:17.941614125 +0000 UTC m=+19.586483860,LastTimestamp:2026-02-27 19:34:17.941614125 +0000 UTC m=+19.586483860,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 19:34:23 crc kubenswrapper[4839]: > Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.685223 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898317b0d26f69d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:17.941685917 +0000 UTC m=+19.586555652,LastTimestamp:2026-02-27 19:34:17.941685917 +0000 UTC m=+19.586555652,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:23 crc kubenswrapper[4839]: W0227 19:34:23.819830 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 27 19:34:23 crc kubenswrapper[4839]: E0227 19:34:23.819898 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:23 crc kubenswrapper[4839]: I0227 19:34:23.912191 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:24 crc kubenswrapper[4839]: I0227 19:34:24.911958 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:25 crc kubenswrapper[4839]: I0227 19:34:25.910766 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:26 crc kubenswrapper[4839]: I0227 19:34:26.514353 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:26 crc kubenswrapper[4839]: I0227 19:34:26.515995 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:26 crc kubenswrapper[4839]: I0227 19:34:26.516104 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:26 crc kubenswrapper[4839]: I0227 19:34:26.516131 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:26 crc kubenswrapper[4839]: I0227 19:34:26.516180 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:26 crc kubenswrapper[4839]: E0227 19:34:26.521720 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:34:26 crc kubenswrapper[4839]: E0227 19:34:26.521858 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:34:26 crc kubenswrapper[4839]: I0227 19:34:26.911933 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:27 crc kubenswrapper[4839]: W0227 19:34:27.729761 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 27 19:34:27 crc kubenswrapper[4839]: E0227 19:34:27.730481 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.909516 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.940774 4839 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.940887 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.940951 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.941537 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.942762 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.942830 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.942857 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.943784 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 27 19:34:27 crc kubenswrapper[4839]: I0227 19:34:27.944127 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83" gracePeriod=30 Feb 27 19:34:27 crc kubenswrapper[4839]: E0227 19:34:27.948232 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898317b0d25de2d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 19:34:27 crc kubenswrapper[4839]: &Event{ObjectMeta:{kube-controller-manager-crc.1898317b0d25de2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 19:34:27 crc kubenswrapper[4839]: body: Feb 27 19:34:27 crc kubenswrapper[4839]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:17.941614125 +0000 UTC m=+19.586483860,LastTimestamp:2026-02-27 19:34:27.940865642 +0000 UTC m=+29.585735397,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 19:34:27 crc kubenswrapper[4839]: > Feb 27 19:34:27 crc kubenswrapper[4839]: E0227 19:34:27.953243 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898317b0d26f69d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898317b0d26f69d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:17.941685917 +0000 UTC m=+19.586555652,LastTimestamp:2026-02-27 19:34:27.940915414 +0000 UTC m=+29.585785169,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:27 crc kubenswrapper[4839]: E0227 19:34:27.958299 4839 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898317d6157a1e4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:27.944096228 +0000 UTC m=+29.588966043,LastTimestamp:2026-02-27 19:34:27.944096228 +0000 UTC m=+29.588966043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:27 crc kubenswrapper[4839]: W0227 19:34:27.981650 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 27 19:34:27 crc kubenswrapper[4839]: E0227 19:34:27.981746 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:28 crc kubenswrapper[4839]: E0227 19:34:28.065770 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18983176e0bf2401\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176e0bf2401 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.016815105 +0000 UTC m=+1.661684850,LastTimestamp:2026-02-27 19:34:28.05987946 +0000 UTC m=+29.704749205,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:28 crc kubenswrapper[4839]: I0227 19:34:28.119442 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 19:34:28 crc kubenswrapper[4839]: I0227 19:34:28.119935 4839 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83" exitCode=255 Feb 27 19:34:28 crc kubenswrapper[4839]: I0227 19:34:28.119969 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83"} Feb 27 19:34:28 crc kubenswrapper[4839]: E0227 19:34:28.254312 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18983176f0a44901\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176f0a44901 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.283490561 +0000 UTC m=+1.928360296,LastTimestamp:2026-02-27 19:34:28.248084928 +0000 UTC m=+29.892954673,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:28 crc kubenswrapper[4839]: E0227 19:34:28.263773 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18983176f1720705\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18983176f1720705 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:00.296974085 +0000 UTC m=+1.941843830,LastTimestamp:2026-02-27 19:34:28.261657417 +0000 UTC m=+29.906527152,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:28 crc kubenswrapper[4839]: I0227 19:34:28.909255 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:29 crc kubenswrapper[4839]: E0227 19:34:29.031385 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:34:29 crc kubenswrapper[4839]: W0227 19:34:29.086057 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:29 crc kubenswrapper[4839]: E0227 19:34:29.086114 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.126319 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.126834 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97"} Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.126955 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.128118 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.128162 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.128172 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:29 crc kubenswrapper[4839]: I0227 19:34:29.912023 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:30 crc kubenswrapper[4839]: I0227 19:34:30.130118 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:30 crc kubenswrapper[4839]: I0227 19:34:30.132754 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:30 crc kubenswrapper[4839]: I0227 19:34:30.132792 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:30 crc kubenswrapper[4839]: I0227 19:34:30.132805 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:30 crc kubenswrapper[4839]: I0227 19:34:30.951948 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:31 crc kubenswrapper[4839]: I0227 19:34:31.919486 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:31 crc kubenswrapper[4839]: I0227 19:34:31.965420 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:31 crc kubenswrapper[4839]: I0227 19:34:31.967260 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:31 crc kubenswrapper[4839]: I0227 19:34:31.967320 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:31 crc kubenswrapper[4839]: I0227 19:34:31.967345 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:31 crc kubenswrapper[4839]: I0227 19:34:31.968407 4839 scope.go:117] "RemoveContainer" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" Feb 27 19:34:32 crc kubenswrapper[4839]: I0227 19:34:32.911948 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.141754 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.143800 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.146328 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" exitCode=255 Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.146549 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e"} Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.146627 4839 scope.go:117] "RemoveContainer" containerID="c091ed22d019f7591e4e275bcb03812ca2fe91ba77196c801b9a901324cad3cc" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.146932 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.148338 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.148393 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.148411 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.149436 4839 scope.go:117] "RemoveContainer" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" Feb 27 19:34:33 crc kubenswrapper[4839]: E0227 19:34:33.149782 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.522208 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.524060 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.524138 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.524164 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.524204 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:33 crc kubenswrapper[4839]: E0227 19:34:33.529399 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:34:33 crc kubenswrapper[4839]: E0227 19:34:33.529944 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:34:33 crc kubenswrapper[4839]: I0227 19:34:33.911786 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.154119 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.912648 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.939760 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.940062 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.941882 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.941967 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:34 crc kubenswrapper[4839]: I0227 19:34:34.941993 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.220540 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.220812 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.222243 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.222299 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.222317 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.223057 4839 scope.go:117] "RemoveContainer" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" Feb 27 19:34:35 crc kubenswrapper[4839]: E0227 19:34:35.223373 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.912441 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.984433 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.984606 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.985852 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.985914 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:35 crc kubenswrapper[4839]: I0227 19:34:35.985937 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.911549 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.946346 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.946583 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.948170 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.948226 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.948245 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:36 crc kubenswrapper[4839]: I0227 19:34:36.949033 4839 scope.go:117] "RemoveContainer" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" Feb 27 19:34:36 crc kubenswrapper[4839]: E0227 19:34:36.949318 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:37 crc kubenswrapper[4839]: I0227 19:34:37.911260 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:37 crc kubenswrapper[4839]: I0227 19:34:37.939864 4839 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:34:37 crc kubenswrapper[4839]: I0227 19:34:37.939933 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:34:37 crc kubenswrapper[4839]: E0227 19:34:37.946035 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898317b0d25de2d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 19:34:37 crc kubenswrapper[4839]: &Event{ObjectMeta:{kube-controller-manager-crc.1898317b0d25de2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 19:34:37 crc kubenswrapper[4839]: body: Feb 27 19:34:37 crc kubenswrapper[4839]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:17.941614125 +0000 UTC m=+19.586483860,LastTimestamp:2026-02-27 19:34:37.939914858 +0000 UTC m=+39.584784633,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 19:34:37 crc kubenswrapper[4839]: > Feb 27 19:34:37 crc kubenswrapper[4839]: E0227 19:34:37.949740 4839 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898317b0d26f69d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898317b0d26f69d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:34:17.941685917 +0000 UTC m=+19.586555652,LastTimestamp:2026-02-27 19:34:37.939956899 +0000 UTC m=+39.584826644,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:34:38 crc kubenswrapper[4839]: I0227 19:34:38.906760 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:39 crc kubenswrapper[4839]: E0227 19:34:39.031517 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:34:39 crc kubenswrapper[4839]: W0227 19:34:39.579134 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 27 19:34:39 crc kubenswrapper[4839]: E0227 19:34:39.579213 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:39 crc kubenswrapper[4839]: I0227 19:34:39.913136 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:40 crc kubenswrapper[4839]: I0227 19:34:40.530426 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:40 crc kubenswrapper[4839]: I0227 19:34:40.531806 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:40 crc kubenswrapper[4839]: I0227 19:34:40.531867 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:40 crc kubenswrapper[4839]: I0227 19:34:40.531885 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:40 crc kubenswrapper[4839]: I0227 19:34:40.531923 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:40 crc kubenswrapper[4839]: E0227 19:34:40.537290 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:34:40 crc kubenswrapper[4839]: E0227 19:34:40.537474 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:34:40 crc kubenswrapper[4839]: I0227 19:34:40.911951 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:41 crc kubenswrapper[4839]: I0227 19:34:41.913200 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:42 crc kubenswrapper[4839]: I0227 19:34:42.911553 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:43 crc kubenswrapper[4839]: I0227 19:34:43.909608 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.908051 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.945958 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.946231 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.947845 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.947881 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.947895 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:44 crc kubenswrapper[4839]: I0227 19:34:44.951337 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:34:45 crc kubenswrapper[4839]: I0227 19:34:45.188086 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:45 crc kubenswrapper[4839]: I0227 19:34:45.189144 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:45 crc kubenswrapper[4839]: I0227 19:34:45.189304 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:45 crc kubenswrapper[4839]: I0227 19:34:45.189379 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:45 crc kubenswrapper[4839]: W0227 19:34:45.619809 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 27 19:34:45 crc kubenswrapper[4839]: E0227 19:34:45.619910 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:45 crc kubenswrapper[4839]: I0227 19:34:45.908451 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:46 crc kubenswrapper[4839]: I0227 19:34:46.911004 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:47 crc kubenswrapper[4839]: I0227 19:34:47.537563 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:47 crc kubenswrapper[4839]: I0227 19:34:47.538637 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:47 crc kubenswrapper[4839]: I0227 19:34:47.538735 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:47 crc kubenswrapper[4839]: I0227 19:34:47.538749 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:47 crc kubenswrapper[4839]: I0227 19:34:47.538775 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:47 crc kubenswrapper[4839]: E0227 19:34:47.542769 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:34:47 crc kubenswrapper[4839]: E0227 19:34:47.542864 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:34:47 crc kubenswrapper[4839]: I0227 19:34:47.912340 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:48 crc kubenswrapper[4839]: W0227 19:34:48.514591 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 27 19:34:48 crc kubenswrapper[4839]: E0227 19:34:48.514692 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:48 crc kubenswrapper[4839]: I0227 19:34:48.909263 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:48 crc kubenswrapper[4839]: I0227 19:34:48.964401 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:48 crc kubenswrapper[4839]: I0227 19:34:48.965573 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:48 crc kubenswrapper[4839]: I0227 19:34:48.965640 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:48 crc kubenswrapper[4839]: I0227 19:34:48.965660 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:48 crc kubenswrapper[4839]: I0227 19:34:48.966495 4839 scope.go:117] "RemoveContainer" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" Feb 27 19:34:48 crc kubenswrapper[4839]: E0227 19:34:48.966904 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:34:49 crc kubenswrapper[4839]: E0227 19:34:49.031697 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:34:49 crc kubenswrapper[4839]: I0227 19:34:49.911148 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:51 crc kubenswrapper[4839]: I0227 19:34:51.245755 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:51 crc kubenswrapper[4839]: I0227 19:34:51.912837 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:52 crc kubenswrapper[4839]: W0227 19:34:52.009559 4839 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:52 crc kubenswrapper[4839]: E0227 19:34:52.009638 4839 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 19:34:52 crc kubenswrapper[4839]: I0227 19:34:52.912219 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:53 crc kubenswrapper[4839]: I0227 19:34:53.912613 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:54 crc kubenswrapper[4839]: I0227 19:34:54.543159 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:54 crc kubenswrapper[4839]: I0227 19:34:54.544789 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:54 crc kubenswrapper[4839]: I0227 19:34:54.544888 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:54 crc kubenswrapper[4839]: I0227 19:34:54.544927 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:54 crc kubenswrapper[4839]: I0227 19:34:54.544974 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:34:54 crc kubenswrapper[4839]: E0227 19:34:54.550069 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:34:54 crc kubenswrapper[4839]: E0227 19:34:54.550243 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:34:54 crc kubenswrapper[4839]: I0227 19:34:54.910228 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:55 crc kubenswrapper[4839]: I0227 19:34:55.216064 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 19:34:55 crc kubenswrapper[4839]: I0227 19:34:55.216247 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:34:55 crc kubenswrapper[4839]: I0227 19:34:55.217660 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:34:55 crc kubenswrapper[4839]: I0227 19:34:55.217749 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:34:55 crc kubenswrapper[4839]: I0227 19:34:55.217766 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:34:55 crc kubenswrapper[4839]: I0227 19:34:55.911954 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:56 crc kubenswrapper[4839]: I0227 19:34:56.911743 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:57 crc kubenswrapper[4839]: I0227 19:34:57.912448 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:58 crc kubenswrapper[4839]: I0227 19:34:58.911490 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:34:59 crc kubenswrapper[4839]: E0227 19:34:59.031913 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:34:59 crc kubenswrapper[4839]: I0227 19:34:59.911793 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:35:00 crc kubenswrapper[4839]: I0227 19:35:00.911586 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:35:00 crc kubenswrapper[4839]: I0227 19:35:00.964985 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:00 crc kubenswrapper[4839]: I0227 19:35:00.966457 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:00 crc kubenswrapper[4839]: I0227 19:35:00.966530 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:00 crc kubenswrapper[4839]: I0227 19:35:00.966548 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:00 crc kubenswrapper[4839]: I0227 19:35:00.967596 4839 scope.go:117] "RemoveContainer" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.233081 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.550939 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.553568 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.553615 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.553632 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.553697 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:35:01 crc kubenswrapper[4839]: E0227 19:35:01.562728 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 19:35:01 crc kubenswrapper[4839]: E0227 19:35:01.562872 4839 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 19:35:01 crc kubenswrapper[4839]: I0227 19:35:01.911072 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.239408 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.241973 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f"} Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.242154 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.243485 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.243653 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.243814 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:02 crc kubenswrapper[4839]: I0227 19:35:02.912798 4839 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.247353 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.248004 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.249982 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" exitCode=255 Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.250013 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f"} Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.250042 4839 scope.go:117] "RemoveContainer" containerID="a429b0bc3d8cad3f308f46f311c8d9ca775a60ae630758cb6d1464924ffdd67e" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.250173 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.251192 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.251215 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.251224 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.251607 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:03 crc kubenswrapper[4839]: E0227 19:35:03.251762 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.491248 4839 csr.go:261] certificate signing request csr-swknw is approved, waiting to be issued Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.501619 4839 csr.go:257] certificate signing request csr-swknw is issued Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.592397 4839 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 27 19:35:03 crc kubenswrapper[4839]: I0227 19:35:03.783248 4839 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 27 19:35:04 crc kubenswrapper[4839]: I0227 19:35:04.254393 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 19:35:04 crc kubenswrapper[4839]: I0227 19:35:04.503489 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-12 21:02:35.558946878 +0000 UTC Feb 27 19:35:04 crc kubenswrapper[4839]: I0227 19:35:04.503552 4839 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7657h27m31.055401859s for next certificate rotation Feb 27 19:35:05 crc kubenswrapper[4839]: I0227 19:35:05.220408 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:35:05 crc kubenswrapper[4839]: I0227 19:35:05.220720 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:05 crc kubenswrapper[4839]: I0227 19:35:05.222508 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:05 crc kubenswrapper[4839]: I0227 19:35:05.222570 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:05 crc kubenswrapper[4839]: I0227 19:35:05.222587 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:05 crc kubenswrapper[4839]: I0227 19:35:05.223600 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:05 crc kubenswrapper[4839]: E0227 19:35:05.223920 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:35:06 crc kubenswrapper[4839]: I0227 19:35:06.946788 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:35:06 crc kubenswrapper[4839]: I0227 19:35:06.946999 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:06 crc kubenswrapper[4839]: I0227 19:35:06.948375 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:06 crc kubenswrapper[4839]: I0227 19:35:06.948425 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:06 crc kubenswrapper[4839]: I0227 19:35:06.948442 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:06 crc kubenswrapper[4839]: I0227 19:35:06.949326 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:06 crc kubenswrapper[4839]: E0227 19:35:06.949657 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.563201 4839 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.564335 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.564396 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.564419 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.564823 4839 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.578371 4839 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.578629 4839 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.578643 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.583197 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.583238 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.583250 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.583268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.583283 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:08Z","lastTransitionTime":"2026-02-27T19:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.598375 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.609937 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.609989 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.610009 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.610035 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.610054 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:08Z","lastTransitionTime":"2026-02-27T19:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.628715 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.638879 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.638939 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.638961 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.638989 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.639012 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:08Z","lastTransitionTime":"2026-02-27T19:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.656427 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.667376 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.667439 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.667463 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.667494 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:08 crc kubenswrapper[4839]: I0227 19:35:08.667520 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:08Z","lastTransitionTime":"2026-02-27T19:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.683928 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.684039 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.684060 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.785128 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.886000 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:08 crc kubenswrapper[4839]: E0227 19:35:08.986950 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.033055 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.087737 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.187870 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.288616 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.389061 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.490001 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.590479 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.690851 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.791141 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.892019 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:09 crc kubenswrapper[4839]: E0227 19:35:09.992173 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.093267 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.194260 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.294367 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.394519 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.495429 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.596069 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.696225 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.797333 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.898350 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:10 crc kubenswrapper[4839]: E0227 19:35:10.999564 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.100660 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.200786 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.301216 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.401601 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.501969 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.602659 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.702865 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.803998 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:11 crc kubenswrapper[4839]: E0227 19:35:11.904808 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.005532 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.105793 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.207038 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.307232 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.408066 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.509215 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.610371 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.710867 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.811278 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:12 crc kubenswrapper[4839]: E0227 19:35:12.911927 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.012659 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.113141 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.213261 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.313748 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.414175 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.515268 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.616206 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.717062 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.817777 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:13 crc kubenswrapper[4839]: E0227 19:35:13.918641 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.019013 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.119759 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.220509 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.321339 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.421907 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.522480 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.623716 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.723872 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.824477 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:14 crc kubenswrapper[4839]: E0227 19:35:14.925608 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.026543 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.127056 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.228236 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.328777 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.429241 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.530007 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.630509 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.731738 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.832712 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:15 crc kubenswrapper[4839]: E0227 19:35:15.933743 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.034980 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.135601 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.235786 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.336758 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.437772 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.538048 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.639226 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.739411 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.839957 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:16 crc kubenswrapper[4839]: E0227 19:35:16.940108 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.041107 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.141229 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.241989 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.343446 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: I0227 19:35:17.419093 4839 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.444295 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.544863 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.645764 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.746611 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.846825 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:17 crc kubenswrapper[4839]: E0227 19:35:17.946995 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.047653 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.148862 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.250149 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.350845 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.451827 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.552781 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.653262 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.754082 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.855221 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.955772 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.961498 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.961560 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.961579 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.961606 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.961628 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:18Z","lastTransitionTime":"2026-02-27T19:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:18 crc kubenswrapper[4839]: E0227 19:35:18.978806 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.987850 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.987927 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.987951 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.987982 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:18 crc kubenswrapper[4839]: I0227 19:35:18.988005 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:18Z","lastTransitionTime":"2026-02-27T19:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.003080 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.006760 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.006820 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.006839 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.006863 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.006880 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:19Z","lastTransitionTime":"2026-02-27T19:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.021427 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.025230 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.025306 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.025332 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.025363 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:19 crc kubenswrapper[4839]: I0227 19:35:19.025384 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:19Z","lastTransitionTime":"2026-02-27T19:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.033982 4839 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.036384 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.036506 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.036529 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.137348 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.238215 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.338871 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.439514 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.540752 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.641168 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.741546 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.842236 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:19 crc kubenswrapper[4839]: E0227 19:35:19.942693 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.043736 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.144862 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: I0227 19:35:20.230694 4839 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.245297 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.345486 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.445605 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.545797 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.646258 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.746388 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.846579 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:20 crc kubenswrapper[4839]: E0227 19:35:20.947742 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:21 crc kubenswrapper[4839]: E0227 19:35:21.048701 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:21 crc kubenswrapper[4839]: E0227 19:35:21.149583 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:21 crc kubenswrapper[4839]: E0227 19:35:21.249995 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.305593 4839 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 19:35:21 crc kubenswrapper[4839]: E0227 19:35:21.350287 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:21 crc kubenswrapper[4839]: E0227 19:35:21.451399 4839 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.481796 4839 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.554382 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.554421 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.554430 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.554444 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.554454 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:21Z","lastTransitionTime":"2026-02-27T19:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.657170 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.657228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.657249 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.657278 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.657301 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:21Z","lastTransitionTime":"2026-02-27T19:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.759168 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.759201 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.759209 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.759222 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.759234 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:21Z","lastTransitionTime":"2026-02-27T19:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.861655 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.861737 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.861755 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.861777 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.861794 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:21Z","lastTransitionTime":"2026-02-27T19:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.964877 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.964949 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.964971 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.964998 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.965018 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:21Z","lastTransitionTime":"2026-02-27T19:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:21 crc kubenswrapper[4839]: I0227 19:35:21.980752 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:21 crc kubenswrapper[4839]: E0227 19:35:21.981046 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.067392 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.067458 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.067478 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.067501 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.067517 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.170523 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.170574 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.170587 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.170607 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.170623 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.254123 4839 apiserver.go:52] "Watching apiserver" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.258761 4839 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.259281 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-ktkfz","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw","openshift-machine-config-operator/machine-config-daemon-thb8n","openshift-multus/multus-additional-cni-plugins-xgp4c","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/iptables-alerter-4ln5h","openshift-ovn-kubernetes/ovnkube-node-t22j4","openshift-dns/node-resolver-8wlb4","openshift-kube-apiserver/kube-apiserver-crc","openshift-multus/multus-2w9pp","openshift-multus/network-metrics-daemon-m6927","openshift-network-diagnostics/network-check-target-xd92c"] Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.259759 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.259972 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.260064 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.260509 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.260532 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.260567 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.260576 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.260519 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.260631 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.261445 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.261472 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.261999 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.262054 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.262163 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.262537 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.262751 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.262886 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.264461 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.269370 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.269627 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.269854 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.270002 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.270009 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.270238 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.270240 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.270401 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.270483 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.271626 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.271645 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.271812 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.271988 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.272064 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.272420 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.272532 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.272715 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273008 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273066 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273137 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273156 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273165 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273182 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273192 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273241 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273251 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273524 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273607 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273649 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273963 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.273967 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274048 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274076 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274086 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274097 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274109 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274160 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274214 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274319 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274443 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274559 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.274827 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.289699 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.304282 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.304480 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.306702 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.315738 4839 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.318207 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.328863 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.344613 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.359351 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.372449 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.375124 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.375302 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.375397 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.375496 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.375590 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.390896 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.401587 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415113 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415281 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415350 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415376 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415400 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415435 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415453 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415474 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415497 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415514 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415534 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415557 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415580 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415598 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415678 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415705 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415723 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415747 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415765 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415781 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415805 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415825 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415846 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415863 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415879 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415901 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415922 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415943 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415969 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415993 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.415998 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416014 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416072 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416104 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416137 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416161 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416186 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416211 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416205 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416238 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416338 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416423 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416251 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416526 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416616 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416884 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416943 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416968 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.416997 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417138 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417158 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417159 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417223 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417349 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417509 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417688 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417886 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.417997 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.418050 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.418250 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.418500 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.418635 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.418657 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.418731 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.419094 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.419487 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.419591 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.419969 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420070 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420091 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420182 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420100 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420308 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420364 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420402 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420479 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420688 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.420787 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:22.920764806 +0000 UTC m=+84.565634551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420781 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420838 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420915 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.420981 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421119 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421200 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421264 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421381 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421399 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421427 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421489 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421601 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421719 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421793 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421856 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421882 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421914 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.421952 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422123 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422139 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422182 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422206 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422192 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422229 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422205 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422298 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422319 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422368 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422404 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422439 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422448 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422509 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422541 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422742 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422901 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.422965 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.423013 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426238 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426578 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426610 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426620 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426723 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426802 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426840 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426751 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.426858 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427080 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427092 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427136 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427177 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427778 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427809 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427817 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427877 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.427904 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428174 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428303 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428314 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428358 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428384 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428424 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429135 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428952 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429013 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429181 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.428996 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429147 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429162 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429298 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429331 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429362 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429390 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429419 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429506 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429539 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429568 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429548 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429598 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429628 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429663 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429718 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429747 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429775 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429805 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429829 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429837 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429885 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429912 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429936 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.429961 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430116 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430205 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430245 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430271 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430335 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430356 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430381 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430401 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430422 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430426 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430471 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430494 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430516 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430550 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430573 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430593 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430618 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430642 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430695 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430719 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430740 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430764 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430787 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430813 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430837 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430859 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430886 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430910 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430936 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430964 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430985 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431013 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431037 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431059 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431083 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431108 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431130 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431153 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431176 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431201 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431226 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431250 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431277 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431308 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431334 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431359 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431386 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431412 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431437 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431460 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431485 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431511 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431537 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431560 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431584 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431610 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431634 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431658 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431710 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431734 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431767 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431794 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431819 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431842 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431863 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431886 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431910 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431930 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431953 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431974 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431996 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432020 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432043 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432067 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432094 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432118 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432147 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432171 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432197 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432221 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432249 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432276 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432301 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432326 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432353 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432378 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432405 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432428 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432452 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432477 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432504 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432530 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432554 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432577 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432602 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432625 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432648 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432703 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432729 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432758 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432812 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f59a1827-952c-4b68-97ed-8d457c3cd3bf-proxy-tls\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432841 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f59a1827-952c-4b68-97ed-8d457c3cd3bf-mcd-auth-proxy-config\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432871 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-cnibin\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432904 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432928 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f59a1827-952c-4b68-97ed-8d457c3cd3bf-rootfs\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433159 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-ovn\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433184 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-env-overrides\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433208 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-script-lib\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433231 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68hlp\" (UniqueName: \"kubernetes.io/projected/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-kube-api-access-68hlp\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433255 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-kubelet\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433277 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-daemon-config\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433310 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433336 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e44a6f76-5d63-4471-b5fb-d407ac6870d8-serviceca\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430582 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430596 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430844 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.430928 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431056 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431085 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431218 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431265 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431384 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431468 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431623 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433441 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431752 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.431761 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432001 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432089 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432104 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432162 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432245 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432452 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.432827 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433110 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433127 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433108 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433288 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433350 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433766 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433810 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433815 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433886 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434285 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434298 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434352 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434431 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434516 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434561 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434613 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434727 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434821 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.434935 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435035 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435047 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435082 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435159 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435457 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435476 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435454 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435583 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435603 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435652 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435722 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.435751 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.436057 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.436816 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.437013 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.437306 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.437352 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.437912 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438161 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438371 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438365 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438402 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438555 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438621 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.438666 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.439336 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.439835 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.439844 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.439907 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440218 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440160 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440230 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440515 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440536 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440652 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.439053 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.440900 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441179 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441204 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441250 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441604 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.433362 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5a463cb-3034-49ae-8c71-7bb8aa057a12-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441730 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b5a463cb-3034-49ae-8c71-7bb8aa057a12-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441767 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-systemd-units\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441790 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-netns\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441811 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-node-log\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441830 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-system-cni-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441855 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-slash\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441876 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-ovn-kubernetes\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441898 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-multus-certs\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441920 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c8c19f2-e237-4d41-b593-452ec944b703-cni-binary-copy\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441941 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-k8s-cni-cncf-io\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441961 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441979 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovn-node-metrics-cert\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.441998 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-os-release\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442021 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b5a463cb-3034-49ae-8c71-7bb8aa057a12-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442039 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442049 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442060 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.442100 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442104 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442148 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3acae75b-e15c-4822-9442-3dc0f53bb3c4-hosts-file\") pod \"node-resolver-8wlb4\" (UID: \"3acae75b-e15c-4822-9442-3dc0f53bb3c4\") " pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.442168 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:22.94214441 +0000 UTC m=+84.587014165 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442187 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-systemd\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442208 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-var-lib-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442231 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442253 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442305 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.442319 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442330 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442354 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lbfn\" (UniqueName: \"kubernetes.io/projected/e44a6f76-5d63-4471-b5fb-d407ac6870d8-kube-api-access-4lbfn\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.442374 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:22.942357628 +0000 UTC m=+84.587227383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442401 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7qf2\" (UniqueName: \"kubernetes.io/projected/b5a463cb-3034-49ae-8c71-7bb8aa057a12-kube-api-access-d7qf2\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442435 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-kubelet\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442465 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-config\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442558 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-socket-dir-parent\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442596 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442625 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6c8c19f2-e237-4d41-b593-452ec944b703-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442659 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442719 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.442944 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443022 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42f6r\" (UniqueName: \"kubernetes.io/projected/3acae75b-e15c-4822-9442-3dc0f53bb3c4-kube-api-access-42f6r\") pod \"node-resolver-8wlb4\" (UID: \"3acae75b-e15c-4822-9442-3dc0f53bb3c4\") " pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443056 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-hostroot\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443086 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-log-socket\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443114 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-netd\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443143 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e44a6f76-5d63-4471-b5fb-d407ac6870d8-host\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443189 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-bin\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443220 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-cni-multus\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443249 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-conf-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443275 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-etc-kubernetes\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443303 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q9pd\" (UniqueName: \"kubernetes.io/projected/6c8c19f2-e237-4d41-b593-452ec944b703-kube-api-access-9q9pd\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443338 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443371 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443404 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-etc-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443435 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnsr6\" (UniqueName: \"kubernetes.io/projected/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-kube-api-access-pnsr6\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443466 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjfkk\" (UniqueName: \"kubernetes.io/projected/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-kube-api-access-hjfkk\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443585 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdtd2\" (UniqueName: \"kubernetes.io/projected/f59a1827-952c-4b68-97ed-8d457c3cd3bf-kube-api-access-bdtd2\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443596 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443618 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-cni-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443649 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-cni-bin\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443711 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443744 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-system-cni-dir\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443778 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443809 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-cnibin\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443840 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-os-release\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443872 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443904 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443935 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-cni-binary-copy\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.443964 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-netns\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.444051 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.444525 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.444795 4839 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.445021 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.445875 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.446519 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.446877 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447103 4839 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447192 4839 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447277 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447352 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447432 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447514 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447590 4839 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447714 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447870 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447927 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.447980 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448036 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448090 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448146 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448209 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448261 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448313 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448374 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448429 4839 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448486 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448541 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448598 4839 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448656 4839 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448885 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.448960 4839 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449022 4839 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449075 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449130 4839 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449187 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449242 4839 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449382 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449451 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449510 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449567 4839 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449622 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449702 4839 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449774 4839 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449830 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449886 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.449944 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450003 4839 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450056 4839 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450113 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450172 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450228 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450284 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450352 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450411 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450465 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450517 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450572 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450624 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450699 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450822 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450884 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450937 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.450995 4839 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451048 4839 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451099 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451155 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451212 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451262 4839 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451312 4839 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451361 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451419 4839 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451474 4839 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451529 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451580 4839 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451630 4839 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451720 4839 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.451827 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452072 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452189 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452321 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452433 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452563 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452643 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452727 4839 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452809 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452877 4839 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452950 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.453031 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.453105 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.453215 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.454270 4839 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.454550 4839 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.454822 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.454983 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455109 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455256 4839 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455368 4839 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455476 4839 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.453461 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.452730 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.454612 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455127 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455694 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455634 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456175 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456204 4839 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456225 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.455507 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456247 4839 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456273 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.456273 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456336 4839 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.456356 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456393 4839 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456420 4839 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.456448 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:22.956414987 +0000 UTC m=+84.601284792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.455542 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456143 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.457070 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458279 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.456493 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458461 4839 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458489 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458514 4839 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458540 4839 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458565 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458587 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458609 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458631 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458652 4839 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458755 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458780 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458802 4839 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458822 4839 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458844 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458866 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458888 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458910 4839 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458933 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458955 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.458979 4839 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459001 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459022 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459068 4839 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459090 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459111 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459132 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459153 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459176 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459198 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459219 4839 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459239 4839 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459262 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459284 4839 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459305 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459331 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459353 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459378 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459403 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459425 4839 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459446 4839 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459467 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459488 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459509 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459532 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459554 4839 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459578 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459601 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459621 4839 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459642 4839 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459669 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459911 4839 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.459933 4839 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.457616 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.463033 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.463893 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.468318 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.468561 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.468580 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.468592 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.468632 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:22.968617475 +0000 UTC m=+84.613487210 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.469833 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.470129 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.470725 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.470775 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.470922 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.471448 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.471526 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.471762 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.473210 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.477935 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.478711 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.478758 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.478776 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.478798 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.478817 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.479991 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.480400 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.480457 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.480473 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.480854 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.480925 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.480958 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.481101 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.485868 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.487355 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.487606 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.487744 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.487851 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.488540 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.489093 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.489129 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.490620 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.495261 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.498696 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.507564 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.512107 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.519327 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.521607 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.532345 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.560963 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f59a1827-952c-4b68-97ed-8d457c3cd3bf-rootfs\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561071 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f59a1827-952c-4b68-97ed-8d457c3cd3bf-rootfs\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561131 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-ovn\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561258 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-ovn\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561276 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-env-overrides\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561405 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-script-lib\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561476 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68hlp\" (UniqueName: \"kubernetes.io/projected/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-kube-api-access-68hlp\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561548 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-kubelet\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561636 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-daemon-config\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561751 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e44a6f76-5d63-4471-b5fb-d407ac6870d8-serviceca\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561653 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-kubelet\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561804 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-env-overrides\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.561917 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5a463cb-3034-49ae-8c71-7bb8aa057a12-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562003 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b5a463cb-3034-49ae-8c71-7bb8aa057a12-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562082 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-systemd-units\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562140 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-systemd-units\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562221 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-netns\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562311 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-node-log\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562258 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-netns\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562386 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-node-log\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562363 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-daemon-config\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562432 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5a463cb-3034-49ae-8c71-7bb8aa057a12-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562416 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-system-cni-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562641 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-slash\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562756 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-ovn-kubernetes\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562825 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-multus-certs\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562904 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c8c19f2-e237-4d41-b593-452ec944b703-cni-binary-copy\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562996 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-k8s-cni-cncf-io\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.563160 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.563286 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovn-node-metrics-cert\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.564813 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-os-release\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.564853 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b5a463cb-3034-49ae-8c71-7bb8aa057a12-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.564870 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-os-release\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562590 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-system-cni-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562848 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-script-lib\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.563121 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-k8s-cni-cncf-io\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562858 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-multus-certs\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.563442 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c8c19f2-e237-4d41-b593-452ec944b703-cni-binary-copy\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.564299 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e44a6f76-5d63-4471-b5fb-d407ac6870d8-serviceca\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562701 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-slash\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.563263 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.562830 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-ovn-kubernetes\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565018 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3acae75b-e15c-4822-9442-3dc0f53bb3c4-hosts-file\") pod \"node-resolver-8wlb4\" (UID: \"3acae75b-e15c-4822-9442-3dc0f53bb3c4\") " pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565046 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-systemd\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565087 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-var-lib-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565130 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565179 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565194 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565209 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565275 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lbfn\" (UniqueName: \"kubernetes.io/projected/e44a6f76-5d63-4471-b5fb-d407ac6870d8-kube-api-access-4lbfn\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565294 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7qf2\" (UniqueName: \"kubernetes.io/projected/b5a463cb-3034-49ae-8c71-7bb8aa057a12-kube-api-access-d7qf2\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565332 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-kubelet\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565347 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-config\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565362 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-socket-dir-parent\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565375 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565440 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b5a463cb-3034-49ae-8c71-7bb8aa057a12-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565440 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565462 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565475 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-var-lib-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565481 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-kubelet\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565507 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3acae75b-e15c-4822-9442-3dc0f53bb3c4-hosts-file\") pod \"node-resolver-8wlb4\" (UID: \"3acae75b-e15c-4822-9442-3dc0f53bb3c4\") " pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.565542 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565572 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-systemd\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565600 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-socket-dir-parent\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565622 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6c8c19f2-e237-4d41-b593-452ec944b703-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.565690 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:23.065659477 +0000 UTC m=+84.710529212 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.565999 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42f6r\" (UniqueName: \"kubernetes.io/projected/3acae75b-e15c-4822-9442-3dc0f53bb3c4-kube-api-access-42f6r\") pod \"node-resolver-8wlb4\" (UID: \"3acae75b-e15c-4822-9442-3dc0f53bb3c4\") " pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566042 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-hostroot\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566063 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-log-socket\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566111 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-netd\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566116 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-config\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566135 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e44a6f76-5d63-4471-b5fb-d407ac6870d8-host\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566165 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-log-socket\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566306 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-netd\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566355 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e44a6f76-5d63-4471-b5fb-d407ac6870d8-host\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566387 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-hostroot\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566416 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-bin\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566464 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-cni-multus\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566517 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-conf-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566565 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-etc-kubernetes\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566619 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q9pd\" (UniqueName: \"kubernetes.io/projected/6c8c19f2-e237-4d41-b593-452ec944b703-kube-api-access-9q9pd\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566652 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-conf-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566705 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-bin\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566723 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-cni-multus\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.566777 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-etc-kubernetes\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567073 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-etc-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567144 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnsr6\" (UniqueName: \"kubernetes.io/projected/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-kube-api-access-pnsr6\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567195 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjfkk\" (UniqueName: \"kubernetes.io/projected/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-kube-api-access-hjfkk\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567286 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdtd2\" (UniqueName: \"kubernetes.io/projected/f59a1827-952c-4b68-97ed-8d457c3cd3bf-kube-api-access-bdtd2\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567313 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-cni-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567505 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-cni-bin\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567640 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-system-cni-dir\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567706 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-cnibin\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567741 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-os-release\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567800 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-var-lib-cni-bin\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567861 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-system-cni-dir\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567908 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-cnibin\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567954 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-multus-cni-dir\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.567991 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-os-release\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568047 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568075 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-cni-binary-copy\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568116 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-netns\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568135 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f59a1827-952c-4b68-97ed-8d457c3cd3bf-proxy-tls\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568147 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-etc-openvswitch\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568194 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-host-run-netns\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568313 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6c8c19f2-e237-4d41-b593-452ec944b703-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568575 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f59a1827-952c-4b68-97ed-8d457c3cd3bf-mcd-auth-proxy-config\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568643 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-cnibin\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568836 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-cnibin\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.568943 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c8c19f2-e237-4d41-b593-452ec944b703-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569021 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569043 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569054 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569063 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569072 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569081 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569091 4839 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569099 4839 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569109 4839 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569120 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569129 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569139 4839 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569147 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569156 4839 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569165 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569173 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569181 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569191 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569199 4839 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569207 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569216 4839 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569225 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569233 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569242 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569252 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569261 4839 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569269 4839 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569280 4839 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569289 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569297 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569308 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569318 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569326 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569335 4839 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569345 4839 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569354 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569362 4839 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569370 4839 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569380 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569388 4839 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.569988 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f59a1827-952c-4b68-97ed-8d457c3cd3bf-mcd-auth-proxy-config\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.571620 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovn-node-metrics-cert\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.571632 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b5a463cb-3034-49ae-8c71-7bb8aa057a12-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.582340 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-cni-binary-copy\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.582379 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.582407 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.582418 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.582433 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.582446 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.583263 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lbfn\" (UniqueName: \"kubernetes.io/projected/e44a6f76-5d63-4471-b5fb-d407ac6870d8-kube-api-access-4lbfn\") pod \"node-ca-ktkfz\" (UID: \"e44a6f76-5d63-4471-b5fb-d407ac6870d8\") " pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.584167 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnsr6\" (UniqueName: \"kubernetes.io/projected/050a1135-8d07-4d7c-89fe-f95d8b1e69f4-kube-api-access-pnsr6\") pod \"multus-2w9pp\" (UID: \"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\") " pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.584819 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdtd2\" (UniqueName: \"kubernetes.io/projected/f59a1827-952c-4b68-97ed-8d457c3cd3bf-kube-api-access-bdtd2\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.585750 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68hlp\" (UniqueName: \"kubernetes.io/projected/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-kube-api-access-68hlp\") pod \"ovnkube-node-t22j4\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.585878 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f59a1827-952c-4b68-97ed-8d457c3cd3bf-proxy-tls\") pod \"machine-config-daemon-thb8n\" (UID: \"f59a1827-952c-4b68-97ed-8d457c3cd3bf\") " pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.586146 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q9pd\" (UniqueName: \"kubernetes.io/projected/6c8c19f2-e237-4d41-b593-452ec944b703-kube-api-access-9q9pd\") pod \"multus-additional-cni-plugins-xgp4c\" (UID: \"6c8c19f2-e237-4d41-b593-452ec944b703\") " pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.586983 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7qf2\" (UniqueName: \"kubernetes.io/projected/b5a463cb-3034-49ae-8c71-7bb8aa057a12-kube-api-access-d7qf2\") pod \"ovnkube-control-plane-749d76644c-2twrw\" (UID: \"b5a463cb-3034-49ae-8c71-7bb8aa057a12\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.587780 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42f6r\" (UniqueName: \"kubernetes.io/projected/3acae75b-e15c-4822-9442-3dc0f53bb3c4-kube-api-access-42f6r\") pod \"node-resolver-8wlb4\" (UID: \"3acae75b-e15c-4822-9442-3dc0f53bb3c4\") " pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.592137 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.593212 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjfkk\" (UniqueName: \"kubernetes.io/projected/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-kube-api-access-hjfkk\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.605013 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.618206 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8wlb4" Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.627944 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-bbbd38759968a85774c827ba4ab65b8a1271eb27baa42b8f615b33a10b4a6b6c WatchSource:0}: Error finding container bbbd38759968a85774c827ba4ab65b8a1271eb27baa42b8f615b33a10b4a6b6c: Status 404 returned error can't find the container with id bbbd38759968a85774c827ba4ab65b8a1271eb27baa42b8f615b33a10b4a6b6c Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.631540 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.644200 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ktkfz" Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.648686 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-8464fd8749f30c4c42b697cbf0d061204136dc6275636b2b1c79f15b60c01988 WatchSource:0}: Error finding container 8464fd8749f30c4c42b697cbf0d061204136dc6275636b2b1c79f15b60c01988: Status 404 returned error can't find the container with id 8464fd8749f30c4c42b697cbf0d061204136dc6275636b2b1c79f15b60c01988 Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.655224 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.666773 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.680970 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.682909 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c8c19f2_e237_4d41_b593_452ec944b703.slice/crio-a7cde97a35119a462efd308e4ee8b9b9329295d40050df2026292f4762b00220 WatchSource:0}: Error finding container a7cde97a35119a462efd308e4ee8b9b9329295d40050df2026292f4762b00220: Status 404 returned error can't find the container with id a7cde97a35119a462efd308e4ee8b9b9329295d40050df2026292f4762b00220 Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.684600 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.684637 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.684649 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.684683 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.684695 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.684717 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f65f341_ed5e_4c1c_8e7c_9c9e556c1323.slice/crio-4bfa4dc5110c35782133a32333727c281789874a34ac1d94fc3fb67b3d182de1 WatchSource:0}: Error finding container 4bfa4dc5110c35782133a32333727c281789874a34ac1d94fc3fb67b3d182de1: Status 404 returned error can't find the container with id 4bfa4dc5110c35782133a32333727c281789874a34ac1d94fc3fb67b3d182de1 Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.690570 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2w9pp" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.699705 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.707829 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf59a1827_952c_4b68_97ed_8d457c3cd3bf.slice/crio-573a7ba7a8f40863970d5eed52af94d1981b96feee2559265b8b3c7c8c6f9bc1 WatchSource:0}: Error finding container 573a7ba7a8f40863970d5eed52af94d1981b96feee2559265b8b3c7c8c6f9bc1: Status 404 returned error can't find the container with id 573a7ba7a8f40863970d5eed52af94d1981b96feee2559265b8b3c7c8c6f9bc1 Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.755083 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod050a1135_8d07_4d7c_89fe_f95d8b1e69f4.slice/crio-c6ce87bbb252172e7349545c757c4420192372d7f3e33af0fd90e84d01079174 WatchSource:0}: Error finding container c6ce87bbb252172e7349545c757c4420192372d7f3e33af0fd90e84d01079174: Status 404 returned error can't find the container with id c6ce87bbb252172e7349545c757c4420192372d7f3e33af0fd90e84d01079174 Feb 27 19:35:22 crc kubenswrapper[4839]: W0227 19:35:22.760834 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5a463cb_3034_49ae_8c71_7bb8aa057a12.slice/crio-6085aaf1014d540dbe8e062112b00e08f1d01b8d6fedd059fed74d3983a26c21 WatchSource:0}: Error finding container 6085aaf1014d540dbe8e062112b00e08f1d01b8d6fedd059fed74d3983a26c21: Status 404 returned error can't find the container with id 6085aaf1014d540dbe8e062112b00e08f1d01b8d6fedd059fed74d3983a26c21 Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.787811 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.787872 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.787886 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.787902 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.787928 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.891463 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.891492 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.891500 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.891513 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.891526 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.972388 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.972960 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.974000 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.974076 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.974099 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.974127 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.974155 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974254 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974261 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974301 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974267 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974335 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:23.974323649 +0000 UTC m=+85.619193384 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974348 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:23.97434303 +0000 UTC m=+85.619212765 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974368 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974385 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974394 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974403 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974407 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:23.974389681 +0000 UTC m=+85.619259406 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974422 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:23.974416332 +0000 UTC m=+85.619286067 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:22 crc kubenswrapper[4839]: E0227 19:35:22.974456 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:23.974450923 +0000 UTC m=+85.619320658 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.974498 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.975344 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.976287 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.977047 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.977639 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.978635 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.979272 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.980198 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.980815 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.981942 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.982687 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.983395 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.984305 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.985052 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.986234 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.986754 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.988260 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.989386 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.990343 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.992410 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.993325 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.994615 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.994641 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.994650 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.994826 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.994845 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.994857 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:22Z","lastTransitionTime":"2026-02-27T19:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.995340 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.996583 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.997241 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.998117 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.998707 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 27 19:35:22 crc kubenswrapper[4839]: I0227 19:35:22.999649 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.000219 4839 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.000329 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.001935 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.002880 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.003284 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.004908 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.005949 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.006441 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.007441 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.008195 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.009060 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.009637 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.011024 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.013547 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.014136 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.017419 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.018067 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.019373 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.020100 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.021070 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.021538 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.022037 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.023442 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.024050 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.075452 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.075585 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.075643 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:24.075626073 +0000 UTC m=+85.720495808 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.097882 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.097930 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.097939 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.097957 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.097971 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.200522 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.200566 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.200577 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.200595 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.200606 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.303208 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.303272 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.303290 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.303336 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.303354 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.307837 4839 generic.go:334] "Generic (PLEG): container finished" podID="6c8c19f2-e237-4d41-b593-452ec944b703" containerID="ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268" exitCode=0 Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.307936 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerDied","Data":"ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.307995 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerStarted","Data":"a7cde97a35119a462efd308e4ee8b9b9329295d40050df2026292f4762b00220"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.309406 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.309445 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2d76ec0fc921362080fed7e5d392439c26db5ea60a1c245f8f23ef4086706dab"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.311109 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" exitCode=0 Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.311208 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.311251 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"4bfa4dc5110c35782133a32333727c281789874a34ac1d94fc3fb67b3d182de1"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.313076 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8wlb4" event={"ID":"3acae75b-e15c-4822-9442-3dc0f53bb3c4","Type":"ContainerStarted","Data":"dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.313155 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8wlb4" event={"ID":"3acae75b-e15c-4822-9442-3dc0f53bb3c4","Type":"ContainerStarted","Data":"1a0f80a9020a290627fe0b40d25397be444b46627814526f703533be1abe8d45"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.315964 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" event={"ID":"b5a463cb-3034-49ae-8c71-7bb8aa057a12","Type":"ContainerStarted","Data":"11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.316034 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" event={"ID":"b5a463cb-3034-49ae-8c71-7bb8aa057a12","Type":"ContainerStarted","Data":"66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.316050 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" event={"ID":"b5a463cb-3034-49ae-8c71-7bb8aa057a12","Type":"ContainerStarted","Data":"6085aaf1014d540dbe8e062112b00e08f1d01b8d6fedd059fed74d3983a26c21"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.317084 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ktkfz" event={"ID":"e44a6f76-5d63-4471-b5fb-d407ac6870d8","Type":"ContainerStarted","Data":"13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.317116 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ktkfz" event={"ID":"e44a6f76-5d63-4471-b5fb-d407ac6870d8","Type":"ContainerStarted","Data":"d29017ec9d915da0981c4a244cbcb389abd16e88a2d28a3fc8ecc5cbe62568fd"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.318757 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.318783 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.318793 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"573a7ba7a8f40863970d5eed52af94d1981b96feee2559265b8b3c7c8c6f9bc1"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.320360 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerStarted","Data":"0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.320434 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerStarted","Data":"c6ce87bbb252172e7349545c757c4420192372d7f3e33af0fd90e84d01079174"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.323049 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.323468 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.323523 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.323533 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8464fd8749f30c4c42b697cbf0d061204136dc6275636b2b1c79f15b60c01988"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.325335 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"bbbd38759968a85774c827ba4ab65b8a1271eb27baa42b8f615b33a10b4a6b6c"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.337349 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.346014 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.353867 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.361115 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.367987 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.382204 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.388560 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.409012 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.409140 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.409174 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.409185 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.409202 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.409211 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.418811 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.432501 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.443598 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.456123 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.472630 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.488907 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.503732 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.511706 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.511742 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.511751 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.511765 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.511776 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.520826 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.531983 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.545553 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.559495 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.576467 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.590446 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.606376 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.613791 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.613825 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.613836 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.613851 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.613861 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.619759 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.633462 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.645029 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.667105 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.678082 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.688805 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.706446 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:23Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.715373 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.715423 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.715433 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.715452 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.715464 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.819239 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.819285 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.819300 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.819321 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.819335 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.921099 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.921132 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.921141 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.921154 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.921164 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:23Z","lastTransitionTime":"2026-02-27T19:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.964939 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.964966 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.964988 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.965058 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.965081 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.965173 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.965371 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.965608 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.986802 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.986914 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.986951 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.986970 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987025 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:25.986977889 +0000 UTC m=+87.631847644 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987055 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987100 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:25.987086973 +0000 UTC m=+87.631956708 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: I0227 19:35:23.987116 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987123 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987146 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987158 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987181 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987192 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987200 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:25.987187756 +0000 UTC m=+87.632057581 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987203 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987233 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:25.987225557 +0000 UTC m=+87.632095302 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987243 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:23 crc kubenswrapper[4839]: E0227 19:35:23.987361 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:25.987325761 +0000 UTC m=+87.632195546 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.024848 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.025202 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.025215 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.025233 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.025245 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.088759 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:24 crc kubenswrapper[4839]: E0227 19:35:24.088894 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:24 crc kubenswrapper[4839]: E0227 19:35:24.088958 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:26.088941965 +0000 UTC m=+87.733811700 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.128101 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.128139 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.128150 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.128175 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.128185 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.232647 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.232695 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.232707 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.232723 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.232734 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.329758 4839 generic.go:334] "Generic (PLEG): container finished" podID="6c8c19f2-e237-4d41-b593-452ec944b703" containerID="5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999" exitCode=0 Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.329825 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerDied","Data":"5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335055 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335084 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335112 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335124 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335138 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335148 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335096 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335206 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335225 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.335234 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.346538 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.362347 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.375486 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.393970 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.404742 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.416900 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.430182 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.439588 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.439619 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.439629 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.439643 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.439654 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.442103 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.452502 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.471883 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.487039 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.501632 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.521010 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.533566 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.541543 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.541583 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.541592 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.541605 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.541614 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.549563 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:24Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.643840 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.643880 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.643889 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.643905 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.643914 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.746425 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.746461 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.746469 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.746483 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.746493 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.849090 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.849343 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.849354 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.849369 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.849381 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.951773 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.951803 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.951811 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.951824 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:24 crc kubenswrapper[4839]: I0227 19:35:24.951833 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:24Z","lastTransitionTime":"2026-02-27T19:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.054636 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.054725 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.054743 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.054768 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.054784 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.156753 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.156803 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.156815 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.156833 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.156845 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.259029 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.259066 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.259076 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.259088 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.259097 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.338476 4839 generic.go:334] "Generic (PLEG): container finished" podID="6c8c19f2-e237-4d41-b593-452ec944b703" containerID="c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b" exitCode=0 Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.338564 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerDied","Data":"c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.344493 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.354987 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.361366 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.361394 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.361402 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.361415 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.361424 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.368968 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.381115 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.393770 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.404576 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.417278 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.427500 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.443750 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.454268 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.463473 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.463504 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.463518 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.463533 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.463546 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.469060 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.480825 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.490798 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.508007 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.521954 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.537801 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:25Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.565189 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.565228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.565238 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.565251 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.565260 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.670902 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.670931 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.670941 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.670957 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.670967 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.773480 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.773523 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.773532 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.773548 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.773557 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.875419 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.875454 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.875464 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.875478 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.875490 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.977989 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:25 crc kubenswrapper[4839]: E0227 19:35:25.978323 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.978028 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:25 crc kubenswrapper[4839]: E0227 19:35:25.978387 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.978118 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:25 crc kubenswrapper[4839]: E0227 19:35:25.978453 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.978002 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:25 crc kubenswrapper[4839]: E0227 19:35:25.978491 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.981635 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.981687 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.981699 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.981712 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:25 crc kubenswrapper[4839]: I0227 19:35:25.981722 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:25Z","lastTransitionTime":"2026-02-27T19:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.008495 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.008593 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.008620 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008691 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008693 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:30.008652917 +0000 UTC m=+91.653522652 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008732 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:30.008719709 +0000 UTC m=+91.653589434 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.008753 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.008792 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008953 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008969 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008975 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008980 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008986 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.008990 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.009014 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:30.009006979 +0000 UTC m=+91.653876714 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.009026 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:30.009020739 +0000 UTC m=+91.653890474 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.009053 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.009086 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:30.009076671 +0000 UTC m=+91.653946406 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.083916 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.083949 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.083956 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.083971 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.083980 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.110102 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.110274 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: E0227 19:35:26.110338 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:30.110316653 +0000 UTC m=+91.755186388 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.186629 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.186722 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.186740 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.186763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.186779 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.289228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.289294 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.289312 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.289338 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.289357 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.350832 4839 generic.go:334] "Generic (PLEG): container finished" podID="6c8c19f2-e237-4d41-b593-452ec944b703" containerID="33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19" exitCode=0 Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.351298 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerDied","Data":"33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.352475 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.367711 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.388175 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.398972 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.399012 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.399021 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.399035 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.399044 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.408894 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.424517 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.443729 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.463016 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.474050 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.493055 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.504234 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.504285 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.504297 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.504314 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.504325 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.504334 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.517472 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.530249 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.540272 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.550950 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.564918 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.578922 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.591507 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.603841 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.606190 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.606210 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.606219 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.606231 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.606241 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.613433 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.628792 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.640244 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.655884 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.671101 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.684740 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.696263 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.708414 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.708457 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.708465 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.708480 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.708493 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.713619 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.726243 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.735959 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.746870 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.756927 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.767230 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:26Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.811132 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.811164 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.811172 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.811185 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.811193 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.913322 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.913359 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.913367 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.913382 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.913392 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:26Z","lastTransitionTime":"2026-02-27T19:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:26 crc kubenswrapper[4839]: I0227 19:35:26.973639 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.015838 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.016064 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.016159 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.016243 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.016311 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.119000 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.119280 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.119381 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.119480 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.119566 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.222204 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.222237 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.222246 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.222258 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.222267 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.325151 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.325211 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.325222 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.325244 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.325256 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.358818 4839 generic.go:334] "Generic (PLEG): container finished" podID="6c8c19f2-e237-4d41-b593-452ec944b703" containerID="4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92" exitCode=0 Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.358883 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerDied","Data":"4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.365591 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.378229 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.393146 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.402221 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.426455 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.427324 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.427363 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.427371 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.427384 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.427393 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.438745 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.454967 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.467951 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.480449 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.494330 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.506843 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.530191 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.530235 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.530251 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.530272 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.530291 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.539637 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.565591 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.582471 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.600269 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.610064 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.618973 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:27Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.632507 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.632541 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.632549 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.632563 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.632573 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.734237 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.734285 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.734302 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.734318 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.734330 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.836865 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.836938 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.836962 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.836996 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.837022 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.939203 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.939265 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.939288 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.939314 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.939337 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:27Z","lastTransitionTime":"2026-02-27T19:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.965202 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.965269 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.965290 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:27 crc kubenswrapper[4839]: I0227 19:35:27.965205 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:27 crc kubenswrapper[4839]: E0227 19:35:27.965470 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:27 crc kubenswrapper[4839]: E0227 19:35:27.965601 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:27 crc kubenswrapper[4839]: E0227 19:35:27.965777 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:27 crc kubenswrapper[4839]: E0227 19:35:27.965929 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.042193 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.042229 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.042244 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.042258 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.042267 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.144321 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.144354 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.144367 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.144381 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.144390 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.246206 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.246260 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.246282 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.246302 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.246317 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.349070 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.349149 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.349168 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.349197 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.349223 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.381940 4839 generic.go:334] "Generic (PLEG): container finished" podID="6c8c19f2-e237-4d41-b593-452ec944b703" containerID="a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f" exitCode=0 Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.382025 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerDied","Data":"a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.399625 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.418124 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.434261 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.453462 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.453490 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.453499 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.453514 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.453525 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.457380 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.467368 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.481725 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.493289 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.505253 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.515473 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.523556 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.546036 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.555479 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.555519 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.555533 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.555550 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.555562 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.558091 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.569444 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.582782 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.597979 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.613231 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.657409 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.657451 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.657462 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.657480 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.657494 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.759725 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.759753 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.759762 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.759775 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.759786 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.863643 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.863733 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.863746 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.863763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.863778 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.966767 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.966821 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.966837 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.966862 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.966877 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:28Z","lastTransitionTime":"2026-02-27T19:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.983213 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:28 crc kubenswrapper[4839]: I0227 19:35:28.994765 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.010274 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.022182 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.034934 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.048601 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.057686 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.068615 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.068701 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.068724 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.068747 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.068762 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.069064 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.081599 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.093425 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.104165 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.115618 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.129053 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.137876 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.153122 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.167061 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.170900 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.170958 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.170971 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.170990 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.171028 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.273790 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.273830 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.273843 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.273860 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.273870 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.358810 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.358868 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.358879 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.358896 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.358908 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.373165 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.377444 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.377475 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.377483 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.377514 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.377523 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.388571 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.389621 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.389680 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.389740 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.395841 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" event={"ID":"6c8c19f2-e237-4d41-b593-452ec944b703","Type":"ContainerStarted","Data":"436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589"} Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.398869 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.404033 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.404074 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.404087 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.404107 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.404119 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.406785 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.414343 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.418914 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.419626 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.423140 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.423191 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.423208 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.423230 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.423254 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.425400 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.437193 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.437342 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.443736 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.443773 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.443785 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.443826 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.443840 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.451813 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.458021 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.458238 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.461947 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.461985 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.461997 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.462025 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.462037 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.463285 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.476952 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.489702 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.501413 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.512826 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.525847 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.541774 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.553895 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.564037 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.564088 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.564097 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.564112 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.564137 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.566891 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.578012 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.589028 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.600916 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.614305 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.626727 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.639234 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.658291 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.666620 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.666653 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.666678 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.666691 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.666700 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.673413 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.687138 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.697270 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.706752 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.715451 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.726584 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.739783 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.755714 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.768318 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.768497 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.768599 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.768693 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.768789 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.770329 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.781783 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.798969 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.811259 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.871118 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.871169 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.871180 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.871199 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.871212 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.965033 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.965059 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.965121 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.965232 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.965304 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.965409 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.965841 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:29 crc kubenswrapper[4839]: E0227 19:35:29.966011 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.973767 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.973794 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.973802 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.973815 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:29 crc kubenswrapper[4839]: I0227 19:35:29.973827 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:29Z","lastTransitionTime":"2026-02-27T19:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.051729 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.051889 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:38.051862409 +0000 UTC m=+99.696732144 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.051978 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.052051 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.052072 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.052102 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052124 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052148 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052162 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052221 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:38.052202951 +0000 UTC m=+99.697072686 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052225 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052243 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052254 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052290 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:38.052278593 +0000 UTC m=+99.697148328 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052289 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052334 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052388 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:38.052364816 +0000 UTC m=+99.697234631 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.052422 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:38.052400227 +0000 UTC m=+99.697270042 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.076793 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.076837 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.076845 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.076860 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.076870 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.153030 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.153248 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: E0227 19:35:30.153359 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:38.153332529 +0000 UTC m=+99.798202304 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.179555 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.179594 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.179605 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.179623 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.179634 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.282703 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.282744 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.282758 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.282777 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.282790 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.385199 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.385242 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.385254 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.385269 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.385282 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.487377 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.487420 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.487432 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.487450 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.487461 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.589725 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.589761 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.589770 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.589783 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.589792 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.691979 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.692016 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.692025 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.692039 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.692050 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.794033 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.794072 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.794082 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.794096 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.794107 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.895945 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.895986 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.895994 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.896008 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.896021 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.998625 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.998664 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.998691 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.998708 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:30 crc kubenswrapper[4839]: I0227 19:35:30.998719 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:30Z","lastTransitionTime":"2026-02-27T19:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.101484 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.101513 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.101520 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.101532 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.101540 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.204115 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.204159 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.204176 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.204195 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.204207 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.307066 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.307102 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.307111 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.307127 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.307139 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.409129 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.409170 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.409180 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.409196 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.409207 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.511551 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.511586 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.511595 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.511609 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.511617 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.613929 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.613966 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.613981 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.613998 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.614008 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.716644 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.716703 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.716715 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.716731 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.716742 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.818802 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.818855 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.818872 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.818895 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.818912 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.921833 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.921885 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.921902 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.921925 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.921941 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:31Z","lastTransitionTime":"2026-02-27T19:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.965157 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.965238 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.965249 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:31 crc kubenswrapper[4839]: E0227 19:35:31.965369 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:31 crc kubenswrapper[4839]: I0227 19:35:31.965381 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:31 crc kubenswrapper[4839]: E0227 19:35:31.965482 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:31 crc kubenswrapper[4839]: E0227 19:35:31.965531 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:31 crc kubenswrapper[4839]: E0227 19:35:31.965609 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.025167 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.025221 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.025239 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.025262 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.025286 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.128135 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.128497 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.128599 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.128709 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.128792 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.232717 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.233044 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.233494 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.233759 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.233957 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.337253 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.337613 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.337774 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.337985 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.338170 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.406170 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/0.log" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.410226 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a" exitCode=1 Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.410288 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.411349 4839 scope.go:117] "RemoveContainer" containerID="1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.439317 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.441847 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.441912 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.441932 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.441959 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.441981 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.466345 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.480118 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.492919 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.504256 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.518736 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.532200 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.543507 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.544698 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.544715 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.544723 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.544736 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.544745 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.554624 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.562973 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.580992 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:31Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 19:35:31.518244 6667 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 19:35:31.518276 6667 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 19:35:31.518351 6667 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 19:35:31.518747 6667 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:31.518768 6667 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:31.518781 6667 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:31.518787 6667 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:31.518800 6667 factory.go:656] Stopping watch factory\\\\nI0227 19:35:31.518798 6667 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:31.518813 6667 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:31.518816 6667 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:31.518828 6667 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.590906 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.601408 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.611605 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.622712 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.632530 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:32Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.647425 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.647461 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.647470 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.647484 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.647494 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.749511 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.749719 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.749727 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.749739 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.749747 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.852066 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.852294 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.852368 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.852428 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.852482 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.955403 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.955657 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.955758 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.955864 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:32 crc kubenswrapper[4839]: I0227 19:35:32.955933 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:32Z","lastTransitionTime":"2026-02-27T19:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.058082 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.058125 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.058137 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.058153 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.058164 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.161259 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.161521 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.161638 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.161757 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.161818 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.264299 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.264493 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.264548 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.264605 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.264676 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.367478 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.367551 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.367566 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.367587 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.367599 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.414071 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/1.log" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.414977 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/0.log" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.421750 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2" exitCode=1 Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.421803 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.421850 4839 scope.go:117] "RemoveContainer" containerID="1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.422922 4839 scope.go:117] "RemoveContainer" containerID="b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2" Feb 27 19:35:33 crc kubenswrapper[4839]: E0227 19:35:33.423266 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.442352 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:31Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 19:35:31.518244 6667 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 19:35:31.518276 6667 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 19:35:31.518351 6667 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 19:35:31.518747 6667 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:31.518768 6667 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:31.518781 6667 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:31.518787 6667 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:31.518800 6667 factory.go:656] Stopping watch factory\\\\nI0227 19:35:31.518798 6667 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:31.518813 6667 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:31.518816 6667 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:31.518828 6667 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:33Z\\\",\\\"message\\\":\\\" 6785 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:33.200097 6785 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:33.200129 6785 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:33.200165 6785 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 19:35:33.200133 6785 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 19:35:33.200133 6785 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 19:35:33.200301 6785 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 19:35:33.200353 6785 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 19:35:33.200357 6785 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:33.200438 6785 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:33.200465 6785 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 19:35:33.200473 6785 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 19:35:33.200481 6785 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:33.200583 6785 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:33.200632 6785 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 19:35:33.200634 6785 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.451826 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.460319 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.470312 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.470349 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.470361 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.470377 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.470390 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.471481 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.482745 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.498486 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.510997 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.523204 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.535426 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.550570 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.564930 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.573404 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.573445 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.573457 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.573475 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.573487 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.577060 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.597511 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.613612 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.623995 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.635755 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:33Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.675544 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.675576 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.675584 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.675597 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.675606 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.777846 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.777900 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.777918 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.777942 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.777960 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.880059 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.880100 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.880114 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.880132 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.880144 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.965358 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.965417 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.965443 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.965365 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:33 crc kubenswrapper[4839]: E0227 19:35:33.965563 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:33 crc kubenswrapper[4839]: E0227 19:35:33.965687 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:33 crc kubenswrapper[4839]: E0227 19:35:33.965862 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:33 crc kubenswrapper[4839]: E0227 19:35:33.966012 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.983196 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.983282 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.983335 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.983416 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:33 crc kubenswrapper[4839]: I0227 19:35:33.983433 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:33Z","lastTransitionTime":"2026-02-27T19:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.085847 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.085906 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.085919 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.085934 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.085944 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.188518 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.188604 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.188618 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.188633 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.188646 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.292165 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.292265 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.292283 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.292313 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.292335 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.394530 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.394604 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.394622 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.394648 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.394696 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.427505 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/1.log" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.497321 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.497359 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.497367 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.497380 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.497389 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.599643 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.599750 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.599769 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.599799 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.599827 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.702467 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.702529 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.702549 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.702611 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.702633 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.804373 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.804411 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.804421 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.804436 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.804446 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.907299 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.907357 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.907376 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.907399 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:34 crc kubenswrapper[4839]: I0227 19:35:34.907418 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:34Z","lastTransitionTime":"2026-02-27T19:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.010290 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.010373 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.010391 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.010417 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.010441 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.113388 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.113459 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.113476 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.113502 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.113520 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.217076 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.217146 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.217171 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.217200 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.217225 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.319731 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.319771 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.319780 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.319794 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.319805 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.422837 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.422905 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.422929 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.422956 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.422973 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.525441 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.525501 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.525517 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.525540 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.525557 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.627467 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.627501 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.627510 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.627522 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.627531 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.731233 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.731283 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.731292 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.731308 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.731317 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.835151 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.835194 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.835215 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.835232 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.835242 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.937442 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.937473 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.937482 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.937495 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.937503 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:35Z","lastTransitionTime":"2026-02-27T19:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.964846 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.964948 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:35 crc kubenswrapper[4839]: E0227 19:35:35.965114 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.965244 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:35 crc kubenswrapper[4839]: E0227 19:35:35.965339 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.965432 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:35 crc kubenswrapper[4839]: E0227 19:35:35.965434 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:35 crc kubenswrapper[4839]: E0227 19:35:35.965736 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:35 crc kubenswrapper[4839]: I0227 19:35:35.965863 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:35 crc kubenswrapper[4839]: E0227 19:35:35.965983 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.040113 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.040660 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.040717 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.040735 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.040746 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.142821 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.142918 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.142939 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.142961 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.142977 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.246296 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.246332 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.246344 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.246360 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.246384 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.348977 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.349085 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.349115 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.349156 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.349183 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.466172 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.466205 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.466215 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.466230 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.466240 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.569084 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.569120 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.569128 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.569142 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.569151 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.672101 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.672172 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.672184 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.672202 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.672216 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.775644 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.775697 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.775710 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.775728 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.775741 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.878957 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.879011 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.879025 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.879045 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.879062 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.981228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.981340 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.981376 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.981402 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:36 crc kubenswrapper[4839]: I0227 19:35:36.981425 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:36Z","lastTransitionTime":"2026-02-27T19:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.084073 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.084131 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.084149 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.084171 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.084189 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.187055 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.187119 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.187138 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.187162 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.187179 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.289348 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.289384 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.289393 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.289407 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.289417 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.392421 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.392481 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.392499 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.392522 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.392539 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.495520 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.495971 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.496127 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.496274 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.496410 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.598415 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.598489 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.598508 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.598538 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.598557 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.701447 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.701513 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.701525 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.701548 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.701563 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.803725 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.803778 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.803794 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.803815 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.803830 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.906900 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.906998 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.907023 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.907059 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.907084 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:37Z","lastTransitionTime":"2026-02-27T19:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.964511 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.964541 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.964634 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:37 crc kubenswrapper[4839]: E0227 19:35:37.964714 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:37 crc kubenswrapper[4839]: I0227 19:35:37.964734 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:37 crc kubenswrapper[4839]: E0227 19:35:37.964843 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:37 crc kubenswrapper[4839]: E0227 19:35:37.964921 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:37 crc kubenswrapper[4839]: E0227 19:35:37.965028 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.010588 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.010648 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.010692 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.010715 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.010732 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.113628 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.113844 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.113856 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.113874 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.113887 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.136532 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.136776 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.136844 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.136927 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:54.136840344 +0000 UTC m=+115.781710119 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.136983 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.137013 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137034 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137069 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137076 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:54.137044561 +0000 UTC m=+115.781914396 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137084 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.137148 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137184 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137243 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:54.137209807 +0000 UTC m=+115.782079642 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137284 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:54.137264598 +0000 UTC m=+115.782134513 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137334 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137370 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137393 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.137473 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:35:54.137452185 +0000 UTC m=+115.782321970 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.217243 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.217290 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.217301 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.217318 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.217329 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.238082 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.238262 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: E0227 19:35:38.238325 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:35:54.238306544 +0000 UTC m=+115.883176289 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.320295 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.320611 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.320799 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.320984 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.321186 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.424180 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.424221 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.424232 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.424252 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.424264 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.526218 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.526268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.526282 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.526300 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.526313 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.628211 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.628264 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.628284 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.628306 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.628324 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.731196 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.731248 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.731264 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.731286 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.731303 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.834555 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.834624 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.834641 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.834710 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.834728 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.937877 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.937941 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.937958 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.937985 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.938003 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:38Z","lastTransitionTime":"2026-02-27T19:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:38 crc kubenswrapper[4839]: I0227 19:35:38.983247 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:38Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.001114 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:38Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.014356 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.026109 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.040623 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.040691 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.040704 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.040721 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.040733 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.049508 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.065947 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.079091 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.092784 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.104098 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.122227 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1794aeaabd1f9f49f272ef885c082c9cba1515d35948c1239b50d3905289f57a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:31Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 19:35:31.518244 6667 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 19:35:31.518276 6667 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 19:35:31.518351 6667 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 19:35:31.518747 6667 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:31.518768 6667 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:31.518781 6667 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:31.518787 6667 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:31.518800 6667 factory.go:656] Stopping watch factory\\\\nI0227 19:35:31.518798 6667 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:31.518813 6667 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:31.518816 6667 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:31.518828 6667 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:33Z\\\",\\\"message\\\":\\\" 6785 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:33.200097 6785 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:33.200129 6785 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:33.200165 6785 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 19:35:33.200133 6785 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 19:35:33.200133 6785 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 19:35:33.200301 6785 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 19:35:33.200353 6785 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 19:35:33.200357 6785 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:33.200438 6785 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:33.200465 6785 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 19:35:33.200473 6785 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 19:35:33.200481 6785 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:33.200583 6785 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:33.200632 6785 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 19:35:33.200634 6785 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.140441 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.143233 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.143267 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.143276 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.143321 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.143332 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.155576 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.167878 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.182466 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.194859 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.209485 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.245562 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.245620 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.245637 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.245661 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.245716 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.348701 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.349024 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.349158 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.349279 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.349380 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.451203 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.451246 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.451259 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.451274 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.451292 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.553066 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.553274 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.553337 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.553398 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.553454 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.648908 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.648982 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.649006 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.649035 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.649059 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.670468 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.675323 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.675380 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.675398 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.675423 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.675441 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.697518 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.702046 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.702113 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.702136 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.702166 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.702189 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.724052 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.729281 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.729450 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.729572 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.729745 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.729843 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.755061 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.760220 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.760275 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.760287 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.760305 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.760320 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.775560 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:39Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.775800 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.777992 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.778061 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.778085 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.778111 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.778129 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.881924 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.881985 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.882004 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.882028 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.882045 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.965144 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.965201 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.965172 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.965152 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.965344 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.965422 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.965470 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:39 crc kubenswrapper[4839]: E0227 19:35:39.965607 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.984740 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.984818 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.984843 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.984873 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:39 crc kubenswrapper[4839]: I0227 19:35:39.984900 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:39Z","lastTransitionTime":"2026-02-27T19:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.088796 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.088851 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.088875 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.088900 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.088918 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.192251 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.192303 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.192320 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.192337 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.192351 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.295020 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.295074 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.295094 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.295116 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.295134 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.397577 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.397642 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.397694 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.397724 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.397745 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.500801 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.500849 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.500861 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.500876 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.501061 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.605110 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.605171 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.605188 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.605214 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.605233 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.708334 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.708383 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.708399 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.708421 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.708472 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.811885 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.811940 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.811956 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.811975 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.811988 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.915520 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.915573 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.915592 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.915614 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.915632 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:40Z","lastTransitionTime":"2026-02-27T19:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:40 crc kubenswrapper[4839]: I0227 19:35:40.987614 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.019130 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.019194 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.019213 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.019237 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.019255 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.122296 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.122347 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.122365 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.122388 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.122405 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.225866 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.225921 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.225939 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.226055 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.226087 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.328763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.328826 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.328844 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.328872 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.328891 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.431355 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.431421 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.431444 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.431474 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.431494 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.535070 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.535134 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.535152 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.535175 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.535191 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.638111 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.638156 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.638168 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.638184 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.638194 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.740435 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.740483 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.740493 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.740508 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.740519 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.843577 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.843647 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.843716 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.843749 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.843769 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.946137 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.946210 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.946228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.946254 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.946272 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:41Z","lastTransitionTime":"2026-02-27T19:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.964858 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.964898 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.964941 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:41 crc kubenswrapper[4839]: I0227 19:35:41.964945 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:41 crc kubenswrapper[4839]: E0227 19:35:41.965044 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:41 crc kubenswrapper[4839]: E0227 19:35:41.965184 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:41 crc kubenswrapper[4839]: E0227 19:35:41.965389 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:41 crc kubenswrapper[4839]: E0227 19:35:41.965567 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.048912 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.048964 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.048980 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.049000 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.049016 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.152619 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.152713 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.152738 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.152767 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.152791 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.255875 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.255940 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.255965 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.255993 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.256019 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.359082 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.359154 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.359173 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.359196 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.359213 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.461476 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.461516 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.461530 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.461546 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.461557 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.564707 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.564739 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.564748 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.564761 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.564770 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.668521 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.668597 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.668619 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.668657 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.668731 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.771306 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.771355 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.771364 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.771380 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.771391 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.875595 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.875722 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.875744 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.875778 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.875813 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.978386 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.978446 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.978470 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.978495 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:42 crc kubenswrapper[4839]: I0227 19:35:42.978516 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:42Z","lastTransitionTime":"2026-02-27T19:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.081560 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.081638 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.081695 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.081730 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.081758 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.185041 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.185109 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.185127 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.185156 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.185173 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.288562 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.288641 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.288656 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.288715 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.288728 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.392181 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.392270 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.392293 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.392324 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.392346 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.495792 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.495833 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.495843 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.495858 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.495868 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.597765 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.597812 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.597824 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.597840 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.597851 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.700785 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.700831 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.700843 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.700862 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.700876 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.803647 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.803700 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.803711 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.803724 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.803733 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.905598 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.905639 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.905647 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.905675 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.905686 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:43Z","lastTransitionTime":"2026-02-27T19:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.965018 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.965082 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.965088 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:43 crc kubenswrapper[4839]: I0227 19:35:43.965156 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:43 crc kubenswrapper[4839]: E0227 19:35:43.965158 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:43 crc kubenswrapper[4839]: E0227 19:35:43.965250 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:43 crc kubenswrapper[4839]: E0227 19:35:43.965353 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:43 crc kubenswrapper[4839]: E0227 19:35:43.965438 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.007887 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.007945 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.007957 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.007973 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.007985 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.110688 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.110732 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.110742 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.110759 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.110773 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.213713 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.213763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.213774 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.213789 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.213801 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.316474 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.316507 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.316515 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.316533 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.316543 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.419387 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.419427 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.419436 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.419450 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.419461 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.522303 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.522342 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.522352 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.522365 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.522374 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.624366 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.624414 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.624429 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.624448 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.624459 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.727409 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.727487 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.727511 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.727544 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.727569 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.830347 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.830393 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.830402 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.830419 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.830430 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.933897 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.933990 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.934020 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.934059 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:44 crc kubenswrapper[4839]: I0227 19:35:44.934088 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:44Z","lastTransitionTime":"2026-02-27T19:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.037271 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.037318 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.037328 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.037345 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.037356 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.139798 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.139844 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.139860 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.139882 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.139899 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.242716 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.242776 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.242790 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.242812 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.242829 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.345179 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.345252 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.345273 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.345300 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.345321 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.447752 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.447814 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.447835 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.447898 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.447915 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.551298 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.551347 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.551362 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.551379 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.551391 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.653763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.653808 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.653818 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.653836 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.653849 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.756089 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.756114 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.756122 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.756159 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.756168 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.859532 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.859577 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.859589 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.859605 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.859616 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.961804 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.962068 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.962134 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.962222 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.962307 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:45Z","lastTransitionTime":"2026-02-27T19:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.965158 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.965229 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.965284 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:45 crc kubenswrapper[4839]: I0227 19:35:45.965369 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:45 crc kubenswrapper[4839]: E0227 19:35:45.965408 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:45 crc kubenswrapper[4839]: E0227 19:35:45.965544 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:45 crc kubenswrapper[4839]: E0227 19:35:45.965487 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:45 crc kubenswrapper[4839]: E0227 19:35:45.965614 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.064839 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.064873 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.064884 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.064902 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.064914 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.166838 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.167105 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.167193 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.167268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.167399 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.270762 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.270816 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.270833 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.270857 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.270875 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.374434 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.374495 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.374514 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.374540 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.374557 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.476637 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.476702 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.476713 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.476729 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.476741 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.580404 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.580452 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.580469 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.580492 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.580507 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.683409 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.683463 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.683481 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.683504 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.683520 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.785917 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.785953 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.785961 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.785975 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.785984 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.888096 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.888145 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.888163 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.888197 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.888233 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.990065 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.990107 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.990117 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.990132 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:46 crc kubenswrapper[4839]: I0227 19:35:46.990145 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:46Z","lastTransitionTime":"2026-02-27T19:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.092442 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.092485 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.092496 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.092516 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.092528 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.195054 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.195099 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.195109 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.195129 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.195140 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.297448 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.297506 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.297521 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.297543 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.297558 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.400555 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.400601 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.400611 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.400626 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.400637 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.504277 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.504326 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.504344 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.504366 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.504387 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.607961 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.608030 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.608049 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.608075 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.608096 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.710968 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.711022 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.711032 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.711072 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.711086 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.813118 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.813171 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.813182 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.813199 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.813231 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.915285 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.915333 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.915345 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.915362 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.915374 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:47Z","lastTransitionTime":"2026-02-27T19:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.964511 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.964543 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.964814 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.964864 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:47 crc kubenswrapper[4839]: E0227 19:35:47.965086 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.965176 4839 scope.go:117] "RemoveContainer" containerID="b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2" Feb 27 19:35:47 crc kubenswrapper[4839]: E0227 19:35:47.965222 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:47 crc kubenswrapper[4839]: E0227 19:35:47.965351 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:47 crc kubenswrapper[4839]: E0227 19:35:47.965401 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.981362 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:47Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:47 crc kubenswrapper[4839]: I0227 19:35:47.995468 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:47Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.012374 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.023192 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.023250 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.023261 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.023284 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.023298 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.024928 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.044361 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.058382 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.077247 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:33Z\\\",\\\"message\\\":\\\" 6785 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:33.200097 6785 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:33.200129 6785 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:33.200165 6785 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 19:35:33.200133 6785 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 19:35:33.200133 6785 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 19:35:33.200301 6785 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 19:35:33.200353 6785 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 19:35:33.200357 6785 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:33.200438 6785 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:33.200465 6785 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 19:35:33.200473 6785 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 19:35:33.200481 6785 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:33.200583 6785 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:33.200632 6785 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 19:35:33.200634 6785 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.087695 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.109322 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.122402 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.126923 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.126998 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.127016 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.127050 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.127070 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.134715 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.145597 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.158804 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.171290 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.183270 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.194448 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.207166 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.230591 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.230654 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.230698 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.230733 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.230754 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.333487 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.333524 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.333539 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.333588 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.333602 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.459670 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.459742 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.459753 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.459770 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.459781 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.483665 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/1.log" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.486236 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.486798 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.506221 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:33Z\\\",\\\"message\\\":\\\" 6785 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:33.200097 6785 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:33.200129 6785 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:33.200165 6785 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 19:35:33.200133 6785 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 19:35:33.200133 6785 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 19:35:33.200301 6785 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 19:35:33.200353 6785 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 19:35:33.200357 6785 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:33.200438 6785 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:33.200465 6785 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 19:35:33.200473 6785 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 19:35:33.200481 6785 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:33.200583 6785 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:33.200632 6785 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 19:35:33.200634 6785 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.517360 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.539467 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.549725 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.561575 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.561620 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.561631 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.561646 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.561658 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.562460 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.572616 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.587575 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.599436 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.613259 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.626413 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.638830 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.651620 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.663814 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.663862 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.663873 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.663892 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.663903 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.665279 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.675537 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.686442 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.696598 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.709331 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.767230 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.767278 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.767291 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.767310 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.767322 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.870001 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.870042 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.870052 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.870069 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.870078 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.975182 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.975233 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.975262 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.975288 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.975305 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:48Z","lastTransitionTime":"2026-02-27T19:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:48 crc kubenswrapper[4839]: I0227 19:35:48.998889 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:48Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.013167 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.025110 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.035994 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.056339 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:33Z\\\",\\\"message\\\":\\\" 6785 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:33.200097 6785 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:33.200129 6785 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:33.200165 6785 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 19:35:33.200133 6785 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 19:35:33.200133 6785 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 19:35:33.200301 6785 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 19:35:33.200353 6785 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 19:35:33.200357 6785 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:33.200438 6785 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:33.200465 6785 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 19:35:33.200473 6785 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 19:35:33.200481 6785 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:33.200583 6785 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:33.200632 6785 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 19:35:33.200634 6785 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.071491 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.078414 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.078451 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.078468 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.078483 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.078495 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.088005 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.100312 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.116208 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.130198 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.142306 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.157243 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.168859 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.179698 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.182358 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.182395 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.182407 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.182423 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.182436 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.190606 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.203913 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.213655 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.285881 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.285948 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.285971 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.286000 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.286024 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.388252 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.388281 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.388289 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.388302 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.388312 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.490210 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.490259 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.490276 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.490298 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.490316 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.492087 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/2.log" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.492820 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/1.log" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.496126 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136" exitCode=1 Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.496186 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.496375 4839 scope.go:117] "RemoveContainer" containerID="b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.497295 4839 scope.go:117] "RemoveContainer" containerID="d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136" Feb 27 19:35:49 crc kubenswrapper[4839]: E0227 19:35:49.497632 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.511922 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.527109 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.539731 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.552137 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.565762 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.581084 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.592257 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.592307 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.592321 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.592343 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.592357 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.592571 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.604620 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.617902 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.628217 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.643719 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.667702 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.679412 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.697574 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.697617 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.697641 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.697666 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.697701 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.708676 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.725535 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.746993 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b627d4120ac679844fd84d2567fc287c7213ecbf4663f1404b8569ef84ad5fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:33Z\\\",\\\"message\\\":\\\" 6785 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 19:35:33.200097 6785 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 19:35:33.200129 6785 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 19:35:33.200165 6785 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 19:35:33.200133 6785 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 19:35:33.200133 6785 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 19:35:33.200301 6785 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 19:35:33.200353 6785 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 19:35:33.200357 6785 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 19:35:33.200438 6785 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 19:35:33.200465 6785 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 19:35:33.200473 6785 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 19:35:33.200481 6785 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 19:35:33.200583 6785 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 19:35:33.200632 6785 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 19:35:33.200634 6785 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.758714 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:49Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.800387 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.800427 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.800436 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.800450 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.800459 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.903138 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.903188 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.903204 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.903228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.903242 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:49Z","lastTransitionTime":"2026-02-27T19:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.964936 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.965092 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:49 crc kubenswrapper[4839]: E0227 19:35:49.965156 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.965186 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:49 crc kubenswrapper[4839]: E0227 19:35:49.965304 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:49 crc kubenswrapper[4839]: I0227 19:35:49.965371 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:49 crc kubenswrapper[4839]: E0227 19:35:49.965437 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:49 crc kubenswrapper[4839]: E0227 19:35:49.965642 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.006216 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.006251 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.006265 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.006280 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.006290 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.109200 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.109261 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.109281 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.109307 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.109325 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.116544 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.116598 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.116615 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.116637 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.116655 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.141543 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.146719 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.146786 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.146814 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.146837 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.146855 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.166960 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.170972 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.171027 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.171045 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.171068 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.171086 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.192784 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.198739 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.198799 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.198816 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.198843 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.198866 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.216482 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.221324 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.221382 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.221403 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.221428 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.221443 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.239478 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.239780 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.242157 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.242198 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.242217 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.242240 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.242257 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.345400 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.345470 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.345489 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.345512 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.345530 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.447955 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.448045 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.448068 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.448091 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.448107 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.503185 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/2.log" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.508576 4839 scope.go:117] "RemoveContainer" containerID="d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136" Feb 27 19:35:50 crc kubenswrapper[4839]: E0227 19:35:50.508851 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.531255 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.551186 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.551247 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.551268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.551292 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.551309 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.553850 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.574252 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.590949 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.611483 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.634577 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.653975 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.654039 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.654055 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.654081 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.654098 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.658621 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.673717 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.690940 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.707582 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.722758 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.748786 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.761431 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.761888 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.761909 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.761927 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.761938 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.767967 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.794978 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.808715 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.823001 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.835204 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:50Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.864003 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.864054 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.864069 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.864086 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.864098 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.966020 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.967371 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.967410 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.967423 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.967438 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:50 crc kubenswrapper[4839]: I0227 19:35:50.967450 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:50Z","lastTransitionTime":"2026-02-27T19:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.070332 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.070370 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.070379 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.070396 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.070405 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.173178 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.173213 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.173224 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.173241 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.173252 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.275129 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.275165 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.275175 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.275191 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.275202 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.377556 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.377611 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.377622 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.377636 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.377647 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.480949 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.481009 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.481025 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.481050 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.481067 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.511740 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.513588 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.513901 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.534068 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.557903 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.580917 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.583536 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.583583 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.583599 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.583623 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.583640 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.598731 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.614795 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.638206 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.658366 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.676098 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.686720 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.686868 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.686891 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.686911 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.686928 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.694615 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.713307 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.730276 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.747850 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.761845 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.776013 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.786959 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.789373 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.789406 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.789415 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.789428 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.789436 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.801641 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.814621 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:51Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.891913 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.891969 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.891988 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.892013 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.892030 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.965159 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.965218 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.965240 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.965173 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:51 crc kubenswrapper[4839]: E0227 19:35:51.965388 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:51 crc kubenswrapper[4839]: E0227 19:35:51.965317 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:51 crc kubenswrapper[4839]: E0227 19:35:51.965455 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:51 crc kubenswrapper[4839]: E0227 19:35:51.965494 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.994474 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.994507 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.994517 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.994530 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:51 crc kubenswrapper[4839]: I0227 19:35:51.994540 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:51Z","lastTransitionTime":"2026-02-27T19:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.097029 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.097079 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.097089 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.097107 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.097118 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.199974 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.200011 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.200019 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.200034 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.200043 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.303039 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.303082 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.303090 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.303103 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.303112 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.404824 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.404849 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.404857 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.404871 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.404880 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.507834 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.507932 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.507953 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.507980 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.508000 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.610363 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.610410 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.610431 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.610461 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.610480 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.714102 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.714139 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.714156 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.714178 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.714195 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.817320 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.817346 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.817358 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.817373 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.817384 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.919930 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.919980 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.919996 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.920018 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:52 crc kubenswrapper[4839]: I0227 19:35:52.920035 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:52Z","lastTransitionTime":"2026-02-27T19:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.022695 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.022743 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.022751 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.022763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.022772 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.125616 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.125733 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.125763 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.125792 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.125816 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.230053 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.230148 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.230173 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.230247 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.230274 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.333193 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.333240 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.333251 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.333268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.333282 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.437060 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.437122 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.437138 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.437162 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.437179 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.540086 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.540632 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.540753 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.540777 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.540796 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.643826 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.643896 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.643914 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.643945 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.643969 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.747321 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.747377 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.747397 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.747423 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.747442 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.850915 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.851009 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.851038 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.851075 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.851101 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.954570 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.954637 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.954696 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.954734 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.954758 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:53Z","lastTransitionTime":"2026-02-27T19:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.965116 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:53 crc kubenswrapper[4839]: E0227 19:35:53.965275 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.965731 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:53 crc kubenswrapper[4839]: E0227 19:35:53.965872 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.965945 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:53 crc kubenswrapper[4839]: E0227 19:35:53.966029 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:53 crc kubenswrapper[4839]: I0227 19:35:53.966093 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:53 crc kubenswrapper[4839]: E0227 19:35:53.966212 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.056914 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.056959 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.056971 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.056988 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.057000 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.159540 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.159600 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.159617 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.159644 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.159698 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.175258 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175448 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:36:26.175410744 +0000 UTC m=+147.820280509 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.175561 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.175649 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.175753 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175836 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175878 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175888 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.175837 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175978 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:36:26.175954292 +0000 UTC m=+147.820824057 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175989 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.175898 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.176007 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.176017 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.176086 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:36:26.176062636 +0000 UTC m=+147.820932411 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.176101 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.176114 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:36:26.176100937 +0000 UTC m=+147.820970712 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.176196 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:36:26.176171859 +0000 UTC m=+147.821041634 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.262784 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.262820 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.262830 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.262845 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.262855 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.277130 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.277252 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: E0227 19:35:54.277317 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:36:26.277300978 +0000 UTC m=+147.922170713 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.365031 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.365073 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.365081 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.365095 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.365105 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.467914 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.467946 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.467956 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.467969 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.467978 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.570611 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.570684 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.570693 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.570736 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.570745 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.674529 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.674586 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.674595 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.674611 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.674622 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.777185 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.777253 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.777271 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.777295 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.777314 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.879579 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.879622 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.879632 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.879645 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.879654 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.982185 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.982268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.982285 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.982325 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:54 crc kubenswrapper[4839]: I0227 19:35:54.982339 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:54Z","lastTransitionTime":"2026-02-27T19:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.084498 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.084540 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.084548 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.084563 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.084574 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.188229 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.188268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.188276 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.188291 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.188300 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.290748 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.290813 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.290831 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.290857 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.290874 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.394997 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.395074 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.395094 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.395123 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.395143 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.497756 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.497826 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.497840 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.497862 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.497876 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.601185 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.601236 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.601248 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.601266 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.601277 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.703239 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.703271 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.703279 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.703292 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.703301 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.806764 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.806817 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.806831 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.806849 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.806864 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.909649 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.909714 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.909725 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.909743 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.909759 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:55Z","lastTransitionTime":"2026-02-27T19:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.964716 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.964793 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.964833 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:55 crc kubenswrapper[4839]: E0227 19:35:55.964952 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:55 crc kubenswrapper[4839]: I0227 19:35:55.964965 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:55 crc kubenswrapper[4839]: E0227 19:35:55.965089 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:55 crc kubenswrapper[4839]: E0227 19:35:55.965200 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:55 crc kubenswrapper[4839]: E0227 19:35:55.965347 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.012343 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.012419 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.012440 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.012471 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.012494 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.116011 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.116076 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.116087 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.116104 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.116116 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.218949 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.218993 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.219002 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.219015 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.219024 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.321523 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.321559 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.321568 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.321582 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.321591 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.424887 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.425022 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.425061 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.425097 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.425122 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.528599 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.528700 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.528715 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.528743 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.528758 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.631543 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.631595 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.631615 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.631641 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.631658 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.735204 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.735257 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.735268 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.735288 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.735301 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.838409 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.838484 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.838521 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.838553 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.838574 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.941122 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.941191 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.941208 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.941228 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:56 crc kubenswrapper[4839]: I0227 19:35:56.941242 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:56Z","lastTransitionTime":"2026-02-27T19:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.044882 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.044937 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.044956 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.044981 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.044998 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.148204 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.148303 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.148318 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.148351 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.148382 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.252822 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.252935 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.252954 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.252977 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.252995 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.355721 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.355762 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.355774 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.355790 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.355800 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.459630 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.459732 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.459757 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.459788 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.459810 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.562184 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.562242 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.562258 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.562281 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.562297 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.666010 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.666066 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.666080 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.666101 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.666115 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.768894 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.768949 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.768968 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.768992 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.769010 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.871698 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.871771 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.871789 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.871811 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.871828 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.964961 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.965003 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.965088 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:57 crc kubenswrapper[4839]: E0227 19:35:57.965095 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.965232 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:57 crc kubenswrapper[4839]: E0227 19:35:57.965353 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:57 crc kubenswrapper[4839]: E0227 19:35:57.965474 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:35:57 crc kubenswrapper[4839]: E0227 19:35:57.965594 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.974168 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.974234 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.974290 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.974316 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:57 crc kubenswrapper[4839]: I0227 19:35:57.974336 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:57Z","lastTransitionTime":"2026-02-27T19:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.077337 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.077389 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.077418 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.077435 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.077447 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.179939 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.179999 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.180015 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.180033 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.180048 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.282806 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.282877 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.282895 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.282921 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.282952 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.385790 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.385850 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.385864 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.385885 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.385900 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.489488 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.489545 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.489562 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.489580 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.489593 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.591955 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.592043 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.592068 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.592098 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.592118 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.694619 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.694711 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.694731 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.694755 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.694776 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.798224 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.798291 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.798311 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.798336 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.798358 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.900534 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.900607 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.900631 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.900660 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.900716 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:35:58Z","lastTransitionTime":"2026-02-27T19:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:35:58 crc kubenswrapper[4839]: I0227 19:35:58.991027 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:58Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: E0227 19:35:59.001847 4839 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.011739 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.028174 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.047051 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: E0227 19:35:59.048720 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.063976 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.079756 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.095398 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.117432 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.136455 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.160153 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.172604 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.183114 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.192138 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.208444 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.226571 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.241063 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.258480 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:35:59Z is after 2025-08-24T17:21:41Z" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.965478 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.965518 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.965621 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:35:59 crc kubenswrapper[4839]: E0227 19:35:59.965699 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:35:59 crc kubenswrapper[4839]: I0227 19:35:59.965796 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:35:59 crc kubenswrapper[4839]: E0227 19:35:59.966011 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:35:59 crc kubenswrapper[4839]: E0227 19:35:59.966082 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:35:59 crc kubenswrapper[4839]: E0227 19:35:59.966219 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.347354 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.347434 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.347453 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.347484 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.347503 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:00Z","lastTransitionTime":"2026-02-27T19:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:00 crc kubenswrapper[4839]: E0227 19:36:00.364742 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:00Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.369944 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.370000 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.370014 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.370036 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.370051 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:00Z","lastTransitionTime":"2026-02-27T19:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:00 crc kubenswrapper[4839]: E0227 19:36:00.390325 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:00Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.395627 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.395725 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.395752 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.395782 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.395807 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:00Z","lastTransitionTime":"2026-02-27T19:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:00 crc kubenswrapper[4839]: E0227 19:36:00.418791 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:00Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.424581 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.424735 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.424758 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.424780 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.424796 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:00Z","lastTransitionTime":"2026-02-27T19:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:00 crc kubenswrapper[4839]: E0227 19:36:00.443840 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:00Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.448371 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.448418 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.448438 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.448465 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:00 crc kubenswrapper[4839]: I0227 19:36:00.448484 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:00Z","lastTransitionTime":"2026-02-27T19:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:00 crc kubenswrapper[4839]: E0227 19:36:00.475107 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:00Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:00 crc kubenswrapper[4839]: E0227 19:36:00.475369 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:36:01 crc kubenswrapper[4839]: I0227 19:36:01.964481 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:01 crc kubenswrapper[4839]: I0227 19:36:01.964538 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:01 crc kubenswrapper[4839]: I0227 19:36:01.964492 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:01 crc kubenswrapper[4839]: E0227 19:36:01.964724 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:01 crc kubenswrapper[4839]: I0227 19:36:01.965080 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:01 crc kubenswrapper[4839]: E0227 19:36:01.965082 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:01 crc kubenswrapper[4839]: E0227 19:36:01.965278 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:01 crc kubenswrapper[4839]: E0227 19:36:01.965381 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:01 crc kubenswrapper[4839]: I0227 19:36:01.965422 4839 scope.go:117] "RemoveContainer" containerID="d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136" Feb 27 19:36:01 crc kubenswrapper[4839]: E0227 19:36:01.965987 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:36:02 crc kubenswrapper[4839]: I0227 19:36:02.977122 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 27 19:36:03 crc kubenswrapper[4839]: I0227 19:36:03.964959 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:03 crc kubenswrapper[4839]: I0227 19:36:03.964987 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:03 crc kubenswrapper[4839]: I0227 19:36:03.965031 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:03 crc kubenswrapper[4839]: E0227 19:36:03.965107 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:03 crc kubenswrapper[4839]: E0227 19:36:03.965192 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:03 crc kubenswrapper[4839]: I0227 19:36:03.965211 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:03 crc kubenswrapper[4839]: E0227 19:36:03.965346 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:03 crc kubenswrapper[4839]: E0227 19:36:03.965508 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:04 crc kubenswrapper[4839]: E0227 19:36:04.049794 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.226579 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.247934 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.270665 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.286755 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.303210 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.317975 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.334353 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.350022 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.367812 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.387396 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.403855 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.425721 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.441298 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.472106 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.486553 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.501836 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.517942 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.533142 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.547240 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:05Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.965293 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.965336 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.965328 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:05 crc kubenswrapper[4839]: E0227 19:36:05.965479 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:05 crc kubenswrapper[4839]: E0227 19:36:05.965565 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:05 crc kubenswrapper[4839]: I0227 19:36:05.965392 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:05 crc kubenswrapper[4839]: E0227 19:36:05.965745 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:05 crc kubenswrapper[4839]: E0227 19:36:05.965866 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:07 crc kubenswrapper[4839]: I0227 19:36:07.965036 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:07 crc kubenswrapper[4839]: I0227 19:36:07.965156 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:07 crc kubenswrapper[4839]: I0227 19:36:07.965036 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:07 crc kubenswrapper[4839]: E0227 19:36:07.965171 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:07 crc kubenswrapper[4839]: I0227 19:36:07.965058 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:07 crc kubenswrapper[4839]: E0227 19:36:07.965355 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:07 crc kubenswrapper[4839]: E0227 19:36:07.965465 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:07 crc kubenswrapper[4839]: E0227 19:36:07.965609 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:08 crc kubenswrapper[4839]: I0227 19:36:08.979648 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:08Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.000148 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:08Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.009480 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.021191 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.036716 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: E0227 19:36:09.050215 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.052756 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.076558 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.088643 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.100381 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.110538 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.130746 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.142465 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.156867 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.170797 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.184333 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.198281 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.211953 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.228434 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.931060 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/0.log" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.931317 4839 generic.go:334] "Generic (PLEG): container finished" podID="050a1135-8d07-4d7c-89fe-f95d8b1e69f4" containerID="0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d" exitCode=1 Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.931342 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerDied","Data":"0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d"} Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.931693 4839 scope.go:117] "RemoveContainer" containerID="0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.948254 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.962703 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.965147 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:09 crc kubenswrapper[4839]: E0227 19:36:09.965278 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.965415 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.965501 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.965648 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:09 crc kubenswrapper[4839]: E0227 19:36:09.965794 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:09 crc kubenswrapper[4839]: E0227 19:36:09.965980 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:09 crc kubenswrapper[4839]: E0227 19:36:09.966039 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:09 crc kubenswrapper[4839]: I0227 19:36:09.986755 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.001010 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:09Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.013577 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.031506 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.053356 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.064282 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.085577 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.097512 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.121194 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.131656 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.144558 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.156762 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.174726 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.187271 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.197782 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.211024 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"2026-02-27T19:35:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f\\\\n2026-02-27T19:35:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f to /host/opt/cni/bin/\\\\n2026-02-27T19:35:24Z [verbose] multus-daemon started\\\\n2026-02-27T19:35:24Z [verbose] Readiness Indicator file check\\\\n2026-02-27T19:36:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.757402 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.757441 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.757452 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.757467 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.757476 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:10Z","lastTransitionTime":"2026-02-27T19:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:10 crc kubenswrapper[4839]: E0227 19:36:10.771595 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.775865 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.775952 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.775977 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.776008 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.776035 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:10Z","lastTransitionTime":"2026-02-27T19:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:10 crc kubenswrapper[4839]: E0227 19:36:10.792374 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.797317 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.797408 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.797432 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.797463 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.797488 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:10Z","lastTransitionTime":"2026-02-27T19:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:10 crc kubenswrapper[4839]: E0227 19:36:10.815436 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.820099 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.820164 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.820184 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.820208 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.820225 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:10Z","lastTransitionTime":"2026-02-27T19:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:10 crc kubenswrapper[4839]: E0227 19:36:10.840431 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.844792 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.844840 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.844850 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.844868 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.844878 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:10Z","lastTransitionTime":"2026-02-27T19:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:10 crc kubenswrapper[4839]: E0227 19:36:10.857324 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: E0227 19:36:10.857822 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.936649 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/0.log" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.936777 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerStarted","Data":"b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250"} Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.950400 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.966779 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:10 crc kubenswrapper[4839]: I0227 19:36:10.984119 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.000148 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:10Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.015146 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.032368 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.048901 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.064227 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.092703 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.105196 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.125926 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.142993 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.159966 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.178046 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"2026-02-27T19:35:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f\\\\n2026-02-27T19:35:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f to /host/opt/cni/bin/\\\\n2026-02-27T19:35:24Z [verbose] multus-daemon started\\\\n2026-02-27T19:35:24Z [verbose] Readiness Indicator file check\\\\n2026-02-27T19:36:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.194850 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.210135 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.223508 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.246073 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:11Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.964550 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.964596 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.964606 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:11 crc kubenswrapper[4839]: I0227 19:36:11.964566 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:11 crc kubenswrapper[4839]: E0227 19:36:11.964737 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:11 crc kubenswrapper[4839]: E0227 19:36:11.964811 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:11 crc kubenswrapper[4839]: E0227 19:36:11.964875 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:11 crc kubenswrapper[4839]: E0227 19:36:11.964926 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:12 crc kubenswrapper[4839]: I0227 19:36:12.965499 4839 scope.go:117] "RemoveContainer" containerID="d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.948217 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/3.log" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.949590 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/2.log" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.952222 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" exitCode=1 Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.952281 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.952343 4839 scope.go:117] "RemoveContainer" containerID="d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.953740 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:36:13 crc kubenswrapper[4839]: E0227 19:36:13.954079 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.964969 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.965075 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.965159 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:13 crc kubenswrapper[4839]: E0227 19:36:13.965153 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:13 crc kubenswrapper[4839]: E0227 19:36:13.965286 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.965381 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:13 crc kubenswrapper[4839]: E0227 19:36:13.965485 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:13 crc kubenswrapper[4839]: E0227 19:36:13.965704 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:13 crc kubenswrapper[4839]: I0227 19:36:13.993751 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:13Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.013659 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.036386 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.050859 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: E0227 19:36:14.051415 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.087797 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:13Z\\\",\\\"message\\\":\\\"shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:13Z is after 2025-08-24T17:21:41Z]\\\\nI0227 19:36:13.687810 7333 services_controller.go:434] Service openshift-controller-manager/controller-manager retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{controller-manager openshift-controller-manager bec3404d-8a9b-42cf-8577-99faf17d6a73 4118 0 2025-02-23 05:22:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-controller-manager] map[operator.openshift.io/spec-hash:b3b96749ab82e4de02ef6aa9f0e168108d09315e18d73931c12251d267378e74 service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:36:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.104375 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.128655 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.145903 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.167343 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.188572 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"2026-02-27T19:35:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f\\\\n2026-02-27T19:35:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f to /host/opt/cni/bin/\\\\n2026-02-27T19:35:24Z [verbose] multus-daemon started\\\\n2026-02-27T19:35:24Z [verbose] Readiness Indicator file check\\\\n2026-02-27T19:36:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.210475 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.229631 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.243492 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.257011 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.272936 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.287192 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.299053 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.311626 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:14Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:14 crc kubenswrapper[4839]: I0227 19:36:14.969928 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/3.log" Feb 27 19:36:15 crc kubenswrapper[4839]: I0227 19:36:15.965207 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:15 crc kubenswrapper[4839]: I0227 19:36:15.965242 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:15 crc kubenswrapper[4839]: I0227 19:36:15.965204 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:15 crc kubenswrapper[4839]: E0227 19:36:15.965338 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:15 crc kubenswrapper[4839]: E0227 19:36:15.965402 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:15 crc kubenswrapper[4839]: I0227 19:36:15.965220 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:15 crc kubenswrapper[4839]: E0227 19:36:15.965709 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:15 crc kubenswrapper[4839]: E0227 19:36:15.965757 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:17 crc kubenswrapper[4839]: I0227 19:36:17.964435 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:17 crc kubenswrapper[4839]: E0227 19:36:17.964703 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:17 crc kubenswrapper[4839]: I0227 19:36:17.965068 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:17 crc kubenswrapper[4839]: E0227 19:36:17.965213 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:17 crc kubenswrapper[4839]: I0227 19:36:17.965017 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:17 crc kubenswrapper[4839]: E0227 19:36:17.965430 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:17 crc kubenswrapper[4839]: I0227 19:36:17.964469 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:17 crc kubenswrapper[4839]: E0227 19:36:17.966443 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:18 crc kubenswrapper[4839]: I0227 19:36:18.984779 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:18Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.000733 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:18Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.038762 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d109e0d3fb368ae82f9f0b1b8c0d1fbcc59edac6a2d2ea932948476d54133136\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:35:48Z\\\",\\\"message\\\":\\\".217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.176\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:1936, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0227 19:35:48.847920 6984 services_controller.go:444] Built service openshift-ingress/router-internal-default LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0227 19:35:48.847929 6984 services_controller.go:445] Built service openshift-ingress/router-internal-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF0227 19:35:48.848000 6984 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:13Z\\\",\\\"message\\\":\\\"shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:13Z is after 2025-08-24T17:21:41Z]\\\\nI0227 19:36:13.687810 7333 services_controller.go:434] Service openshift-controller-manager/controller-manager retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{controller-manager openshift-controller-manager bec3404d-8a9b-42cf-8577-99faf17d6a73 4118 0 2025-02-23 05:22:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-controller-manager] map[operator.openshift.io/spec-hash:b3b96749ab82e4de02ef6aa9f0e168108d09315e18d73931c12251d267378e74 service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:36:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: E0227 19:36:19.052118 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.063126 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.091365 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.109112 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.160644 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.185357 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"2026-02-27T19:35:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f\\\\n2026-02-27T19:35:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f to /host/opt/cni/bin/\\\\n2026-02-27T19:35:24Z [verbose] multus-daemon started\\\\n2026-02-27T19:35:24Z [verbose] Readiness Indicator file check\\\\n2026-02-27T19:36:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.203125 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.217209 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.232421 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.249084 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.261726 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.277510 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.290148 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.303509 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.314377 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.328778 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:19Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.965368 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.965489 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.965407 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:19 crc kubenswrapper[4839]: I0227 19:36:19.965367 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:19 crc kubenswrapper[4839]: E0227 19:36:19.965596 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:19 crc kubenswrapper[4839]: E0227 19:36:19.965722 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:19 crc kubenswrapper[4839]: E0227 19:36:19.965902 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:19 crc kubenswrapper[4839]: E0227 19:36:19.966084 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.259419 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.259463 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.259478 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.259501 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.259517 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:21Z","lastTransitionTime":"2026-02-27T19:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.274628 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:21Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.278651 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.278720 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.278735 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.278757 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.278770 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:21Z","lastTransitionTime":"2026-02-27T19:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.298630 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:21Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.302745 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.302804 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.302828 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.302858 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.302880 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:21Z","lastTransitionTime":"2026-02-27T19:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.315884 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:21Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.319045 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.319187 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.319257 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.319337 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.319426 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:21Z","lastTransitionTime":"2026-02-27T19:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.331887 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:21Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.335440 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.335476 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.335489 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.335506 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.335518 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:21Z","lastTransitionTime":"2026-02-27T19:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.346807 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:21Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.346964 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.965000 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.965049 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.965614 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.965874 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:21 crc kubenswrapper[4839]: I0227 19:36:21.965906 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.966240 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.966461 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:21 crc kubenswrapper[4839]: E0227 19:36:21.966323 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.667101 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.667919 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:36:22 crc kubenswrapper[4839]: E0227 19:36:22.668077 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.688056 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:13Z\\\",\\\"message\\\":\\\"shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:13Z is after 2025-08-24T17:21:41Z]\\\\nI0227 19:36:13.687810 7333 services_controller.go:434] Service openshift-controller-manager/controller-manager retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{controller-manager openshift-controller-manager bec3404d-8a9b-42cf-8577-99faf17d6a73 4118 0 2025-02-23 05:22:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-controller-manager] map[operator.openshift.io/spec-hash:b3b96749ab82e4de02ef6aa9f0e168108d09315e18d73931c12251d267378e74 service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:36:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.701518 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.731022 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.744396 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.758183 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.768830 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.780736 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.796868 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.815963 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.835398 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"2026-02-27T19:35:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f\\\\n2026-02-27T19:35:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f to /host/opt/cni/bin/\\\\n2026-02-27T19:35:24Z [verbose] multus-daemon started\\\\n2026-02-27T19:35:24Z [verbose] Readiness Indicator file check\\\\n2026-02-27T19:36:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.856318 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.873120 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.890240 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.901460 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.914965 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.924140 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.937490 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.948792 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:22Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:22 crc kubenswrapper[4839]: I0227 19:36:22.975948 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 19:36:23 crc kubenswrapper[4839]: I0227 19:36:23.964546 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:23 crc kubenswrapper[4839]: I0227 19:36:23.964618 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:23 crc kubenswrapper[4839]: I0227 19:36:23.964706 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:23 crc kubenswrapper[4839]: E0227 19:36:23.964716 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:23 crc kubenswrapper[4839]: I0227 19:36:23.964566 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:23 crc kubenswrapper[4839]: E0227 19:36:23.964807 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:23 crc kubenswrapper[4839]: E0227 19:36:23.964887 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:23 crc kubenswrapper[4839]: E0227 19:36:23.965028 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:24 crc kubenswrapper[4839]: E0227 19:36:24.053356 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:25 crc kubenswrapper[4839]: I0227 19:36:25.965223 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:25 crc kubenswrapper[4839]: E0227 19:36:25.965730 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:25 crc kubenswrapper[4839]: I0227 19:36:25.965329 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:25 crc kubenswrapper[4839]: I0227 19:36:25.965286 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:25 crc kubenswrapper[4839]: I0227 19:36:25.965355 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:25 crc kubenswrapper[4839]: E0227 19:36:25.966064 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:25 crc kubenswrapper[4839]: E0227 19:36:25.966147 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:25 crc kubenswrapper[4839]: E0227 19:36:25.965932 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:26 crc kubenswrapper[4839]: I0227 19:36:26.214525 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:36:26 crc kubenswrapper[4839]: I0227 19:36:26.214690 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:26 crc kubenswrapper[4839]: I0227 19:36:26.214745 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:26 crc kubenswrapper[4839]: I0227 19:36:26.214780 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:26 crc kubenswrapper[4839]: I0227 19:36:26.214828 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215069 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215129 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215153 4839 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215172 4839 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215243 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.215212011 +0000 UTC m=+211.860081786 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215284 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.215263952 +0000 UTC m=+211.860133717 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215303 4839 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215367 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.215345145 +0000 UTC m=+211.860214920 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215402 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.215385336 +0000 UTC m=+211.860255111 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215069 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215442 4839 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215463 4839 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.215520 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.21550153 +0000 UTC m=+211.860371305 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 19:36:26 crc kubenswrapper[4839]: I0227 19:36:26.315798 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.316030 4839 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:36:26 crc kubenswrapper[4839]: E0227 19:36:26.316137 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs podName:ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.316112805 +0000 UTC m=+211.960982590 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs") pod "network-metrics-daemon-m6927" (UID: "ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 19:36:27 crc kubenswrapper[4839]: I0227 19:36:27.964722 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:27 crc kubenswrapper[4839]: I0227 19:36:27.964803 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:27 crc kubenswrapper[4839]: I0227 19:36:27.964840 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:27 crc kubenswrapper[4839]: I0227 19:36:27.964840 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:27 crc kubenswrapper[4839]: E0227 19:36:27.964904 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:27 crc kubenswrapper[4839]: E0227 19:36:27.964995 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:27 crc kubenswrapper[4839]: E0227 19:36:27.965062 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:27 crc kubenswrapper[4839]: E0227 19:36:27.965207 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:28 crc kubenswrapper[4839]: I0227 19:36:28.979020 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b47d4e99d8ca87fb909877be22b8d4dbcfa2a55cf67c963f91a11ec28d8f9b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:28 crc kubenswrapper[4839]: I0227 19:36:28.994139 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8c19f2-e237-4d41-b593-452ec944b703\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://436748ba3f3475ff1c96c1f27d32905e65d2cf663dcda4b976e3b9de12640589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea8ff695383a7e0162db3c1c4ee1e0328e09a342fdeff3a564ad0ab9b6bef268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5df66d704c59ba7e3e2e004929360d8f10ef3f3168677bb565542483dccf9999\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a5dd91cdf91ee35b3a97f7921e9a646a4dbb688fe89a5b5a35f9f9cf43c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33f1d20e370a34132659791e98a0bd394c2a6599597eadf0d1ea7a73b9bd9e19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e891164befd61949cdd922034483b621831a9da3a4a38ba722a4b61fbf55f92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3b0e57a3ad4625ebc8bc5edc2e518bf04ce78bd3f9de0c836734b1c2c3c826f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9q9pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xgp4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:28Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.011864 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6927" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjfkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6927\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.027007 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5a463cb-3034-49ae-8c71-7bb8aa057a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e8d46f5385f11a7fdc4d7bfa56ba1c5a48a5b65f93557e7ece84df7b9d155d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b2a405ecddf5b9197eb6597e496055b06ed92870a8609e1260adcc8909c281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7qf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2twrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.040096 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4cf9a8e-1a64-4ece-bc50-a31164c4e80c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b01f939c623f89e0272728eb984ea4615a9b9be1eb8a3cfe7670c8a6f2bb130\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d67640a193eab91481ec3764342bebdf83ac0cb0378dbce0b3971347040a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: E0227 19:36:29.054035 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.060879 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ab00b44-f619-4b1b-957a-642eb93210a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9aa13f82fc046b7795dc8c8b8e2835c2442fc5408aaba928d376194311c0bb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff4e5314dbb7a68a3bf58e9882298cc743113c523e907feb73a803be9f990d83\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:34:27Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 19:34:00.831327 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 19:34:00.833912 1 observer_polling.go:159] Starting file observer\\\\nI0227 19:34:00.866035 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 19:34:00.869724 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0227 19:34:27.948224 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0227 19:34:27.948365 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4927642d281660b53cf0277522a1a7530125f61a0ca598a84e25cbb7f57753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6388feaf7cea6b85cacbbfe42ffb962e908aaa12d1834fcace47da392d657b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.080736 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"784d78ba-52b8-45ac-ad7f-eae27798f9ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7d8881df9fc1481b72c4c1cfc7175b4685e0346664391961b45ca1135fab6a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64633ef809e1b51516a270afe403fcfc5134006ccc054f436d23cfc38071d5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb207eb4a6d2a0dfe3cc465f4a9932bb272534cd99be6d5d31a887b72dfcc062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ebeca656e1b52ae5de19c5e2f8fe7adab3294d03c66bf4c30e7d3d7e2070c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebeca656e1b52ae5de19c5e2f8fe7adab3294d03c66bf4c30e7d3d7e2070c55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.094802 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8wlb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3acae75b-e15c-4822-9442-3dc0f53bb3c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dba28ed82d0aea1d451a9f45e367124146bc93ab4010fb21491b9db18aee8e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42f6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8wlb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.105526 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c579ecf15fa152eea0683e67b39c4d2f89ca61c72cc59194adb759ffec1e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ed7374207635564151594b478778a97351f32fba4821ab7d0ea5e2130da054e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.118288 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f59a1827-952c-4b68-97ed-8d457c3cd3bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fee425156588c65c4a52bbe4a7a59ab001a045bea62fc04c3fbb892c7150c74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdtd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-thb8n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.134504 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e05deb-ff83-40f8-bf80-92bc4ba8d393\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf23ac7aa1fc674bff0b1ff20e08666f895fd37721457e141856202e6de8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0424c962c63ed394199657476b83f1b097406a8f03a1c15505747a9fa60b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96cf7cd893787037829cfc5d9a14495fa5e8f0fdf0ef1f720771b0b31a4466ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5e16a98fb883da1df8cfc58f0697c759b40cc56a681d12c73011f34045034ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fb8935b11dfca13cee2428f0ce86cedd4fdcb87cb47124b7943e5a06bfa1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471042d000193eb0e1b24ca72c715f476647cce110407dea6410d8e3cbfd98ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb0022868fad03246ce2b945d2722108b304c1aba29ef0e08b92d7ed93499ab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0725272f5f81be58a9051c56d0b95d8afd74081625b3a3ac21cb2125f7be105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.148582 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8874275f0514c7a01c9833ece14068d8ecbdf86f5ebae6f53cb8a6334a2055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.159697 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.168322 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ktkfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e44a6f76-5d63-4471-b5fb-d407ac6870d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1841ed0d9e3ae5fec796d9b92d7b120107b7be72d2dba30af9d275bf147a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4lbfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ktkfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.190265 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:13Z\\\",\\\"message\\\":\\\"shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:13Z is after 2025-08-24T17:21:41Z]\\\\nI0227 19:36:13.687810 7333 services_controller.go:434] Service openshift-controller-manager/controller-manager retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{controller-manager openshift-controller-manager bec3404d-8a9b-42cf-8577-99faf17d6a73 4118 0 2025-02-23 05:22:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-controller-manager] map[operator.openshift.io/spec-hash:b3b96749ab82e4de02ef6aa9f0e168108d09315e18d73931c12251d267378e74 service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:36:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68hlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t22j4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.206934 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c04a518-2635-44d2-bff2-1332fd913a23\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T19:35:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW0227 19:35:02.096307 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 19:35:02.096473 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 19:35:02.097571 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4091548855/tls.crt::/tmp/serving-cert-4091548855/tls.key\\\\\\\"\\\\nI0227 19:35:02.339230 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 19:35:02.342580 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 19:35:02.342612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 19:35:02.342652 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 19:35:02.342694 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 19:35:02.349527 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0227 19:35:02.349567 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0227 19:35:02.349582 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349594 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 19:35:02.349605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 19:35:02.349613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 19:35:02.349619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 19:35:02.349624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0227 19:35:02.351858 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:35:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:34:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T19:34:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T19:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:33:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.221761 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.240371 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.259649 4839 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2w9pp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"050a1135-8d07-4d7c-89fe-f95d8b1e69f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T19:36:09Z\\\",\\\"message\\\":\\\"2026-02-27T19:35:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f\\\\n2026-02-27T19:35:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56ec006a-0b60-43ed-a0d6-76df6161027f to /host/opt/cni/bin/\\\\n2026-02-27T19:35:24Z [verbose] multus-daemon started\\\\n2026-02-27T19:35:24Z [verbose] Readiness Indicator file check\\\\n2026-02-27T19:36:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T19:35:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T19:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pnsr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T19:35:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2w9pp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:29Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.965364 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.965434 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.965470 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:29 crc kubenswrapper[4839]: I0227 19:36:29.965520 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:29 crc kubenswrapper[4839]: E0227 19:36:29.965532 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:29 crc kubenswrapper[4839]: E0227 19:36:29.965616 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:29 crc kubenswrapper[4839]: E0227 19:36:29.965795 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:29 crc kubenswrapper[4839]: E0227 19:36:29.965925 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.727000 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.727104 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.727122 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.727150 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.727164 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:31Z","lastTransitionTime":"2026-02-27T19:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.744349 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:31Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.748226 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.748490 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.748607 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.748728 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.748820 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:31Z","lastTransitionTime":"2026-02-27T19:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.760707 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:31Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.764010 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.764140 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.764207 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.764271 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.764333 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:31Z","lastTransitionTime":"2026-02-27T19:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.776285 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:31Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.780346 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.780579 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.780735 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.780890 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.781136 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:31Z","lastTransitionTime":"2026-02-27T19:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.794899 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:31Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.798702 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.798734 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.798744 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.798757 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.798766 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:31Z","lastTransitionTime":"2026-02-27T19:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.812020 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T19:36:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da580aef-40a0-4e5a-b9b4-d01d8f39448d\\\",\\\"systemUUID\\\":\\\"425800eb-b94c-4d28-9446-64bd59d9ebe7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T19:36:31Z is after 2025-08-24T17:21:41Z" Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.812175 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.965170 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.965284 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.965332 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:31 crc kubenswrapper[4839]: I0227 19:36:31.965375 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.965514 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.965799 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.965918 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:31 crc kubenswrapper[4839]: E0227 19:36:31.966018 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:33 crc kubenswrapper[4839]: I0227 19:36:33.965429 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:33 crc kubenswrapper[4839]: I0227 19:36:33.965433 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:33 crc kubenswrapper[4839]: E0227 19:36:33.965726 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:33 crc kubenswrapper[4839]: I0227 19:36:33.965468 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:33 crc kubenswrapper[4839]: I0227 19:36:33.965456 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:33 crc kubenswrapper[4839]: E0227 19:36:33.965897 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:33 crc kubenswrapper[4839]: E0227 19:36:33.965981 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:33 crc kubenswrapper[4839]: E0227 19:36:33.966083 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:34 crc kubenswrapper[4839]: E0227 19:36:34.055633 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:34 crc kubenswrapper[4839]: I0227 19:36:34.965787 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:36:34 crc kubenswrapper[4839]: E0227 19:36:34.966052 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:36:35 crc kubenswrapper[4839]: I0227 19:36:35.964525 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:35 crc kubenswrapper[4839]: I0227 19:36:35.964639 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:35 crc kubenswrapper[4839]: E0227 19:36:35.964791 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:35 crc kubenswrapper[4839]: I0227 19:36:35.964971 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:35 crc kubenswrapper[4839]: I0227 19:36:35.965077 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:35 crc kubenswrapper[4839]: E0227 19:36:35.965305 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:35 crc kubenswrapper[4839]: E0227 19:36:35.965396 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:35 crc kubenswrapper[4839]: E0227 19:36:35.965467 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:37 crc kubenswrapper[4839]: I0227 19:36:37.965341 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:37 crc kubenswrapper[4839]: E0227 19:36:37.965452 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:37 crc kubenswrapper[4839]: I0227 19:36:37.965510 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:37 crc kubenswrapper[4839]: I0227 19:36:37.965599 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:37 crc kubenswrapper[4839]: E0227 19:36:37.965862 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:37 crc kubenswrapper[4839]: E0227 19:36:37.966617 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:37 crc kubenswrapper[4839]: I0227 19:36:37.966777 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:37 crc kubenswrapper[4839]: E0227 19:36:37.966844 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:38 crc kubenswrapper[4839]: I0227 19:36:38.988438 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podStartSLOduration=115.988408775 podStartE2EDuration="1m55.988408775s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:38.987875438 +0000 UTC m=+160.632745183" watchObservedRunningTime="2026-02-27 19:36:38.988408775 +0000 UTC m=+160.633278550" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.035082 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=59.035063688 podStartE2EDuration="59.035063688s" podCreationTimestamp="2026-02-27 19:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.021871444 +0000 UTC m=+160.666741179" watchObservedRunningTime="2026-02-27 19:36:39.035063688 +0000 UTC m=+160.679933423" Feb 27 19:36:39 crc kubenswrapper[4839]: E0227 19:36:39.056129 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.070021 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-ktkfz" podStartSLOduration=117.070005344 podStartE2EDuration="1m57.070005344s" podCreationTimestamp="2026-02-27 19:34:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.069391055 +0000 UTC m=+160.714260820" watchObservedRunningTime="2026-02-27 19:36:39.070005344 +0000 UTC m=+160.714875079" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.117634 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.117617217 podStartE2EDuration="1m18.117617217s" podCreationTimestamp="2026-02-27 19:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.117360149 +0000 UTC m=+160.762229924" watchObservedRunningTime="2026-02-27 19:36:39.117617217 +0000 UTC m=+160.762486952" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.183323 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2w9pp" podStartSLOduration=116.183305477 podStartE2EDuration="1m56.183305477s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.165889581 +0000 UTC m=+160.810759306" watchObservedRunningTime="2026-02-27 19:36:39.183305477 +0000 UTC m=+160.828175222" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.215369 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xgp4c" podStartSLOduration=116.215335532 podStartE2EDuration="1m56.215335532s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.203593264 +0000 UTC m=+160.848463009" watchObservedRunningTime="2026-02-27 19:36:39.215335532 +0000 UTC m=+160.860205317" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.237923 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2twrw" podStartSLOduration=116.23790682 podStartE2EDuration="1m56.23790682s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.224994795 +0000 UTC m=+160.869864550" watchObservedRunningTime="2026-02-27 19:36:39.23790682 +0000 UTC m=+160.882776555" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.238411 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=73.238405516 podStartE2EDuration="1m13.238405516s" podCreationTimestamp="2026-02-27 19:35:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.238062245 +0000 UTC m=+160.882932000" watchObservedRunningTime="2026-02-27 19:36:39.238405516 +0000 UTC m=+160.883275251" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.256562 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=37.256537974 podStartE2EDuration="37.256537974s" podCreationTimestamp="2026-02-27 19:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.255951266 +0000 UTC m=+160.900821011" watchObservedRunningTime="2026-02-27 19:36:39.256537974 +0000 UTC m=+160.901407749" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.270591 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=17.270572755 podStartE2EDuration="17.270572755s" podCreationTimestamp="2026-02-27 19:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.269973496 +0000 UTC m=+160.914843231" watchObservedRunningTime="2026-02-27 19:36:39.270572755 +0000 UTC m=+160.915442490" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.284972 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-8wlb4" podStartSLOduration=117.284955496 podStartE2EDuration="1m57.284955496s" podCreationTimestamp="2026-02-27 19:34:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:39.284549343 +0000 UTC m=+160.929419108" watchObservedRunningTime="2026-02-27 19:36:39.284955496 +0000 UTC m=+160.929825231" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.964894 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.964954 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.964954 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:39 crc kubenswrapper[4839]: E0227 19:36:39.965000 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:39 crc kubenswrapper[4839]: I0227 19:36:39.965054 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:39 crc kubenswrapper[4839]: E0227 19:36:39.965188 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:39 crc kubenswrapper[4839]: E0227 19:36:39.965288 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:39 crc kubenswrapper[4839]: E0227 19:36:39.965327 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:41 crc kubenswrapper[4839]: I0227 19:36:41.965532 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:41 crc kubenswrapper[4839]: I0227 19:36:41.965510 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:41 crc kubenswrapper[4839]: I0227 19:36:41.965539 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:41 crc kubenswrapper[4839]: I0227 19:36:41.965635 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:41 crc kubenswrapper[4839]: E0227 19:36:41.965810 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:41 crc kubenswrapper[4839]: E0227 19:36:41.965975 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:41 crc kubenswrapper[4839]: E0227 19:36:41.966027 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:41 crc kubenswrapper[4839]: E0227 19:36:41.966094 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.021881 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.022169 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.022370 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.022535 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.022725 4839 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T19:36:42Z","lastTransitionTime":"2026-02-27T19:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.080027 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c"] Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.080919 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.083622 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.083748 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.083949 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.084126 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.176726 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.176845 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.176884 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.176915 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.177051 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278249 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278351 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278403 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278407 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278445 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278535 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.278562 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.280057 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.289046 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.303984 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b9d27ee-f3c9-4ebe-9489-a4315d60abd9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zwz6c\" (UID: \"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.304530 4839 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.314560 4839 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 19:36:42 crc kubenswrapper[4839]: I0227 19:36:42.403324 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" Feb 27 19:36:42 crc kubenswrapper[4839]: W0227 19:36:42.421910 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b9d27ee_f3c9_4ebe_9489_a4315d60abd9.slice/crio-53df8690afe8ccc31afb3219d9f5c440d6490b2c5725ea6fe5a0054f9e478cef WatchSource:0}: Error finding container 53df8690afe8ccc31afb3219d9f5c440d6490b2c5725ea6fe5a0054f9e478cef: Status 404 returned error can't find the container with id 53df8690afe8ccc31afb3219d9f5c440d6490b2c5725ea6fe5a0054f9e478cef Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.058280 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" event={"ID":"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9","Type":"ContainerStarted","Data":"6393af1544402f1f135f728df1df1025abc76e1ddf5e09eadd088ca5482515da"} Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.058329 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" event={"ID":"2b9d27ee-f3c9-4ebe-9489-a4315d60abd9","Type":"ContainerStarted","Data":"53df8690afe8ccc31afb3219d9f5c440d6490b2c5725ea6fe5a0054f9e478cef"} Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.075205 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwz6c" podStartSLOduration=120.075188614 podStartE2EDuration="2m0.075188614s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:36:43.074275565 +0000 UTC m=+164.719145340" watchObservedRunningTime="2026-02-27 19:36:43.075188614 +0000 UTC m=+164.720058349" Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.965222 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.965245 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.965245 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:43 crc kubenswrapper[4839]: I0227 19:36:43.965216 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:43 crc kubenswrapper[4839]: E0227 19:36:43.965384 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:43 crc kubenswrapper[4839]: E0227 19:36:43.965457 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:43 crc kubenswrapper[4839]: E0227 19:36:43.965557 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:43 crc kubenswrapper[4839]: E0227 19:36:43.965754 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:44 crc kubenswrapper[4839]: E0227 19:36:44.057850 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:45 crc kubenswrapper[4839]: I0227 19:36:45.965056 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:45 crc kubenswrapper[4839]: I0227 19:36:45.965063 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:45 crc kubenswrapper[4839]: I0227 19:36:45.965176 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:45 crc kubenswrapper[4839]: I0227 19:36:45.965213 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:45 crc kubenswrapper[4839]: E0227 19:36:45.965332 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:45 crc kubenswrapper[4839]: E0227 19:36:45.965413 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:45 crc kubenswrapper[4839]: E0227 19:36:45.965521 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:45 crc kubenswrapper[4839]: E0227 19:36:45.965643 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:46 crc kubenswrapper[4839]: I0227 19:36:46.966777 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:36:46 crc kubenswrapper[4839]: E0227 19:36:46.966990 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t22j4_openshift-ovn-kubernetes(1f65f341-ed5e-4c1c-8e7c-9c9e556c1323)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" Feb 27 19:36:47 crc kubenswrapper[4839]: I0227 19:36:47.965003 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:47 crc kubenswrapper[4839]: I0227 19:36:47.965062 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:47 crc kubenswrapper[4839]: I0227 19:36:47.965146 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:47 crc kubenswrapper[4839]: I0227 19:36:47.965023 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:47 crc kubenswrapper[4839]: E0227 19:36:47.965252 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:47 crc kubenswrapper[4839]: E0227 19:36:47.965384 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:47 crc kubenswrapper[4839]: E0227 19:36:47.965490 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:47 crc kubenswrapper[4839]: E0227 19:36:47.965591 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:49 crc kubenswrapper[4839]: E0227 19:36:49.058572 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:49 crc kubenswrapper[4839]: I0227 19:36:49.964622 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:49 crc kubenswrapper[4839]: E0227 19:36:49.964985 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:49 crc kubenswrapper[4839]: I0227 19:36:49.965202 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:49 crc kubenswrapper[4839]: E0227 19:36:49.965304 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:49 crc kubenswrapper[4839]: I0227 19:36:49.965347 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:49 crc kubenswrapper[4839]: E0227 19:36:49.965495 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:49 crc kubenswrapper[4839]: I0227 19:36:49.965585 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:49 crc kubenswrapper[4839]: E0227 19:36:49.965743 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:51 crc kubenswrapper[4839]: I0227 19:36:51.964869 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:51 crc kubenswrapper[4839]: I0227 19:36:51.964882 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:51 crc kubenswrapper[4839]: I0227 19:36:51.964947 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:51 crc kubenswrapper[4839]: I0227 19:36:51.964991 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:51 crc kubenswrapper[4839]: E0227 19:36:51.965164 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:51 crc kubenswrapper[4839]: E0227 19:36:51.965309 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:51 crc kubenswrapper[4839]: E0227 19:36:51.965385 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:51 crc kubenswrapper[4839]: E0227 19:36:51.965463 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:53 crc kubenswrapper[4839]: I0227 19:36:53.964308 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:53 crc kubenswrapper[4839]: I0227 19:36:53.964374 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:53 crc kubenswrapper[4839]: I0227 19:36:53.964392 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:53 crc kubenswrapper[4839]: I0227 19:36:53.964349 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:53 crc kubenswrapper[4839]: E0227 19:36:53.964517 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:53 crc kubenswrapper[4839]: E0227 19:36:53.964458 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:53 crc kubenswrapper[4839]: E0227 19:36:53.964694 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:53 crc kubenswrapper[4839]: E0227 19:36:53.964718 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:54 crc kubenswrapper[4839]: E0227 19:36:54.060109 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:55 crc kubenswrapper[4839]: I0227 19:36:55.964506 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:55 crc kubenswrapper[4839]: I0227 19:36:55.964560 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:55 crc kubenswrapper[4839]: I0227 19:36:55.964608 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:55 crc kubenswrapper[4839]: I0227 19:36:55.964563 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:55 crc kubenswrapper[4839]: E0227 19:36:55.964826 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:55 crc kubenswrapper[4839]: E0227 19:36:55.965208 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:55 crc kubenswrapper[4839]: E0227 19:36:55.965453 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:55 crc kubenswrapper[4839]: E0227 19:36:55.966066 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:56 crc kubenswrapper[4839]: I0227 19:36:56.227531 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/1.log" Feb 27 19:36:56 crc kubenswrapper[4839]: I0227 19:36:56.228290 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/0.log" Feb 27 19:36:56 crc kubenswrapper[4839]: I0227 19:36:56.228613 4839 generic.go:334] "Generic (PLEG): container finished" podID="050a1135-8d07-4d7c-89fe-f95d8b1e69f4" containerID="b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250" exitCode=1 Feb 27 19:36:56 crc kubenswrapper[4839]: I0227 19:36:56.228736 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerDied","Data":"b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250"} Feb 27 19:36:56 crc kubenswrapper[4839]: I0227 19:36:56.228783 4839 scope.go:117] "RemoveContainer" containerID="0dcf791e13bdbfedc56bc3fc8ed0998226500b0813c833c6a95621be89e5b33d" Feb 27 19:36:56 crc kubenswrapper[4839]: I0227 19:36:56.229797 4839 scope.go:117] "RemoveContainer" containerID="b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250" Feb 27 19:36:56 crc kubenswrapper[4839]: E0227 19:36:56.231760 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2w9pp_openshift-multus(050a1135-8d07-4d7c-89fe-f95d8b1e69f4)\"" pod="openshift-multus/multus-2w9pp" podUID="050a1135-8d07-4d7c-89fe-f95d8b1e69f4" Feb 27 19:36:57 crc kubenswrapper[4839]: I0227 19:36:57.233050 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/1.log" Feb 27 19:36:57 crc kubenswrapper[4839]: I0227 19:36:57.965069 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:57 crc kubenswrapper[4839]: I0227 19:36:57.965176 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:57 crc kubenswrapper[4839]: I0227 19:36:57.965226 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:57 crc kubenswrapper[4839]: E0227 19:36:57.965360 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:57 crc kubenswrapper[4839]: I0227 19:36:57.965501 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:57 crc kubenswrapper[4839]: E0227 19:36:57.965591 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:57 crc kubenswrapper[4839]: E0227 19:36:57.965497 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:57 crc kubenswrapper[4839]: E0227 19:36:57.965750 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:36:59 crc kubenswrapper[4839]: E0227 19:36:59.061130 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:36:59 crc kubenswrapper[4839]: I0227 19:36:59.965204 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:36:59 crc kubenswrapper[4839]: I0227 19:36:59.965273 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:36:59 crc kubenswrapper[4839]: E0227 19:36:59.965385 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:36:59 crc kubenswrapper[4839]: I0227 19:36:59.965420 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:36:59 crc kubenswrapper[4839]: I0227 19:36:59.965517 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:36:59 crc kubenswrapper[4839]: E0227 19:36:59.965595 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:36:59 crc kubenswrapper[4839]: E0227 19:36:59.965782 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:36:59 crc kubenswrapper[4839]: E0227 19:36:59.965952 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:00 crc kubenswrapper[4839]: I0227 19:37:00.967035 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.246435 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/3.log" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.248392 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerStarted","Data":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.248793 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.288428 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podStartSLOduration=138.288411236 podStartE2EDuration="2m18.288411236s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:01.288185329 +0000 UTC m=+182.933055064" watchObservedRunningTime="2026-02-27 19:37:01.288411236 +0000 UTC m=+182.933280971" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.842248 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m6927"] Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.842698 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:01 crc kubenswrapper[4839]: E0227 19:37:01.842818 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.964902 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.964952 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:01 crc kubenswrapper[4839]: I0227 19:37:01.964963 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:01 crc kubenswrapper[4839]: E0227 19:37:01.965072 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:01 crc kubenswrapper[4839]: E0227 19:37:01.965185 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:01 crc kubenswrapper[4839]: E0227 19:37:01.965253 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:03 crc kubenswrapper[4839]: I0227 19:37:03.965079 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:03 crc kubenswrapper[4839]: I0227 19:37:03.965082 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:03 crc kubenswrapper[4839]: E0227 19:37:03.965354 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:03 crc kubenswrapper[4839]: I0227 19:37:03.965153 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:03 crc kubenswrapper[4839]: E0227 19:37:03.965512 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:03 crc kubenswrapper[4839]: I0227 19:37:03.965116 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:03 crc kubenswrapper[4839]: E0227 19:37:03.965621 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:03 crc kubenswrapper[4839]: E0227 19:37:03.965806 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:04 crc kubenswrapper[4839]: E0227 19:37:04.062931 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:37:05 crc kubenswrapper[4839]: I0227 19:37:05.965413 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:05 crc kubenswrapper[4839]: I0227 19:37:05.965442 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:05 crc kubenswrapper[4839]: E0227 19:37:05.965552 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:05 crc kubenswrapper[4839]: I0227 19:37:05.965407 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:05 crc kubenswrapper[4839]: E0227 19:37:05.965714 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:05 crc kubenswrapper[4839]: I0227 19:37:05.965800 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:05 crc kubenswrapper[4839]: E0227 19:37:05.965973 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:05 crc kubenswrapper[4839]: E0227 19:37:05.966120 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:07 crc kubenswrapper[4839]: I0227 19:37:07.964510 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:07 crc kubenswrapper[4839]: I0227 19:37:07.964530 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:07 crc kubenswrapper[4839]: I0227 19:37:07.964564 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:07 crc kubenswrapper[4839]: I0227 19:37:07.964752 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:07 crc kubenswrapper[4839]: E0227 19:37:07.966274 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:07 crc kubenswrapper[4839]: E0227 19:37:07.966401 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:07 crc kubenswrapper[4839]: E0227 19:37:07.966522 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:07 crc kubenswrapper[4839]: E0227 19:37:07.966604 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:09 crc kubenswrapper[4839]: E0227 19:37:09.063326 4839 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 19:37:09 crc kubenswrapper[4839]: I0227 19:37:09.964577 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:09 crc kubenswrapper[4839]: I0227 19:37:09.964710 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:09 crc kubenswrapper[4839]: E0227 19:37:09.964742 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:09 crc kubenswrapper[4839]: I0227 19:37:09.964579 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:09 crc kubenswrapper[4839]: I0227 19:37:09.964601 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:09 crc kubenswrapper[4839]: E0227 19:37:09.964998 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:09 crc kubenswrapper[4839]: E0227 19:37:09.965000 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:09 crc kubenswrapper[4839]: E0227 19:37:09.965401 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:10 crc kubenswrapper[4839]: I0227 19:37:10.965504 4839 scope.go:117] "RemoveContainer" containerID="b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250" Feb 27 19:37:11 crc kubenswrapper[4839]: I0227 19:37:11.293805 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/1.log" Feb 27 19:37:11 crc kubenswrapper[4839]: I0227 19:37:11.294128 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerStarted","Data":"818928ceb7d6dc64a37536581b2018af0ff97985fed3a856d9292c377d5a5d0a"} Feb 27 19:37:11 crc kubenswrapper[4839]: I0227 19:37:11.965365 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:11 crc kubenswrapper[4839]: I0227 19:37:11.965444 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:11 crc kubenswrapper[4839]: E0227 19:37:11.965531 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:11 crc kubenswrapper[4839]: I0227 19:37:11.965585 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:11 crc kubenswrapper[4839]: I0227 19:37:11.965594 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:11 crc kubenswrapper[4839]: E0227 19:37:11.965714 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:11 crc kubenswrapper[4839]: E0227 19:37:11.965888 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:11 crc kubenswrapper[4839]: E0227 19:37:11.965952 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:13 crc kubenswrapper[4839]: I0227 19:37:13.965225 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:13 crc kubenswrapper[4839]: I0227 19:37:13.965329 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:13 crc kubenswrapper[4839]: I0227 19:37:13.965225 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:13 crc kubenswrapper[4839]: E0227 19:37:13.965423 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 19:37:13 crc kubenswrapper[4839]: E0227 19:37:13.965498 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6927" podUID="ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b" Feb 27 19:37:13 crc kubenswrapper[4839]: I0227 19:37:13.965701 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:13 crc kubenswrapper[4839]: E0227 19:37:13.965753 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 19:37:13 crc kubenswrapper[4839]: E0227 19:37:13.965741 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.964804 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.964859 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.964925 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.964827 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.968566 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.968566 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.968719 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.968925 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.968939 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 19:37:15 crc kubenswrapper[4839]: I0227 19:37:15.970639 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.681946 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.682297 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.688339 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.839555 4839 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.885005 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.885817 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.886043 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9kxx2"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.886708 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.886973 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbbmg"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.887440 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.890723 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.890866 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.890957 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.891813 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.892273 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.892388 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.892823 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.892864 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.893095 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.893134 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.895837 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pgbmd"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.895853 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.896458 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.897433 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.898083 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.901117 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.902158 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.904623 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.905069 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v6dfv"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.905517 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.911072 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.913340 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: W0227 19:37:22.913938 4839 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 27 19:37:22 crc kubenswrapper[4839]: E0227 19:37:22.913977 4839 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.914132 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.916781 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.917108 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.917278 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.922381 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.922635 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.922896 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.922919 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923039 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923046 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923157 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923167 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923290 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923346 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923396 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923500 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.923616 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.924103 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.924602 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.924784 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.924969 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.925150 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.925258 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.926142 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.926298 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.926543 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gbl86"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.927139 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.927276 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.927344 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.927914 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.928130 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.928316 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qfgcq"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.929229 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.929348 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.929510 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.941460 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.943553 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.961826 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.965304 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.965444 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.965939 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.966306 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.966393 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.966528 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.967098 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.979072 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pc7k8"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.979591 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.980100 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-82v8f"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.980532 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.981014 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.982803 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.983353 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.983452 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.984056 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.988241 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.988760 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf6mj"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.989116 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.989809 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.990380 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.990717 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.990960 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.991197 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.991256 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.991429 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.991381 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.991907 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.992686 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.992882 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.993071 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.993060 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.993200 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.993371 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.993114 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.993943 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.994445 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.994526 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.994597 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.994698 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.995718 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.995993 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.996529 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.996860 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.997128 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.997308 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.997447 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vgnmn"] Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.997628 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 19:37:22 crc kubenswrapper[4839]: I0227 19:37:22.998008 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.000938 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001225 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001371 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001458 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001540 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001732 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001926 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.001992 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.002101 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.002106 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.002297 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.004229 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qmwh7"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.004895 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fszhl"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.019644 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.020645 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.021129 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.021348 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.021532 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.024003 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.024860 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.049228 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-85jfr"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.050038 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.050065 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.050360 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.051608 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.051641 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.051998 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.053123 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.053234 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.053883 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.055039 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.055587 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.056387 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.056999 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.059364 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.061163 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.061279 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.061288 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.062419 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.058793 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.063593 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.063834 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.064899 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.067955 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068493 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-serving-cert\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068529 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-trusted-ca\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068701 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-trusted-ca-bundle\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068734 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45305c05-bdb6-4b38-acdb-2fb7f472ed68-metrics-tls\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068754 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-config\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068773 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-etcd-client\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068792 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-etcd-serving-ca\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068812 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-config\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068831 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c828db27-a26e-41dd-a68e-d9137c3b25bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068853 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-auth-proxy-config\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068874 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m97gl\" (UniqueName: \"kubernetes.io/projected/b6f71939-65e0-4459-9c58-211f5cafd259-kube-api-access-m97gl\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068895 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068915 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068935 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-config\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068977 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jz5j\" (UniqueName: \"kubernetes.io/projected/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-kube-api-access-9jz5j\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.068996 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-oauth-config\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069017 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-images\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069051 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlmx5\" (UniqueName: \"kubernetes.io/projected/45305c05-bdb6-4b38-acdb-2fb7f472ed68-kube-api-access-vlmx5\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069075 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069094 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-image-import-ca\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069113 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-serving-cert\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069136 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-audit-dir\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069156 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvzgj\" (UniqueName: \"kubernetes.io/projected/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-kube-api-access-hvzgj\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069175 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-machine-approver-tls\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069196 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcf6w\" (UniqueName: \"kubernetes.io/projected/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-kube-api-access-wcf6w\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069220 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wksrn\" (UniqueName: \"kubernetes.io/projected/e105188e-f3d1-4757-8134-0db7ce5cbbf2-kube-api-access-wksrn\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069241 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-audit-policies\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069260 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069279 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-config\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069298 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-audit\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069318 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9tvh\" (UniqueName: \"kubernetes.io/projected/710bae1b-9785-4c51-a128-f91f4e962113-kube-api-access-d9tvh\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069340 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-oauth-serving-cert\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069357 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99hh7\" (UniqueName: \"kubernetes.io/projected/e14522ed-1f77-462c-a5ec-caa1efe74a35-kube-api-access-99hh7\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069380 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-config\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069402 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c376125-89b2-43d4-970a-c13ff4fbed59-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069423 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c376125-89b2-43d4-970a-c13ff4fbed59-config\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069493 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-etcd-client\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069576 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-policies\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069594 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069606 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069646 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-config\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069685 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-service-ca\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069725 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069742 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8b87c1a-1579-4415-8134-021c1d7c390e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069763 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l5dd\" (UniqueName: \"kubernetes.io/projected/1c376125-89b2-43d4-970a-c13ff4fbed59-kube-api-access-6l5dd\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069781 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-serving-cert\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069797 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069813 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-client-ca\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069830 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ef6e180-79df-40ee-bf2b-96332e35c052-serving-cert\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069846 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-encryption-config\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069862 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e14522ed-1f77-462c-a5ec-caa1efe74a35-audit-dir\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069878 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zls89\" (UniqueName: \"kubernetes.io/projected/a4baae9f-93ba-40bf-9c28-01f13cdcaf43-kube-api-access-zls89\") pod \"downloads-7954f5f757-82v8f\" (UID: \"a4baae9f-93ba-40bf-9c28-01f13cdcaf43\") " pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069895 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8b87c1a-1579-4415-8134-021c1d7c390e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069911 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv96b\" (UniqueName: \"kubernetes.io/projected/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-kube-api-access-zv96b\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069926 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069942 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.069966 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070029 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070063 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070083 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/710bae1b-9785-4c51-a128-f91f4e962113-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070118 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4v9v\" (UniqueName: \"kubernetes.io/projected/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-kube-api-access-h4v9v\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070138 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45305c05-bdb6-4b38-acdb-2fb7f472ed68-bound-sa-token\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070156 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c828db27-a26e-41dd-a68e-d9137c3b25bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070182 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f71939-65e0-4459-9c58-211f5cafd259-serving-cert\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070198 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/444f893d-0c15-4a0c-8805-38f51f8037f6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-729pl\" (UID: \"444f893d-0c15-4a0c-8805-38f51f8037f6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070214 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070228 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e14522ed-1f77-462c-a5ec-caa1efe74a35-node-pullsecrets\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070247 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-client-ca\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070261 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070275 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/710bae1b-9785-4c51-a128-f91f4e962113-serving-cert\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070332 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwwsg\" (UniqueName: \"kubernetes.io/projected/f8b87c1a-1579-4415-8134-021c1d7c390e-kube-api-access-xwwsg\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070349 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk262\" (UniqueName: \"kubernetes.io/projected/444f893d-0c15-4a0c-8805-38f51f8037f6-kube-api-access-gk262\") pod \"cluster-samples-operator-665b6dd947-729pl\" (UID: \"444f893d-0c15-4a0c-8805-38f51f8037f6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070366 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070387 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zldkj\" (UniqueName: \"kubernetes.io/projected/c828db27-a26e-41dd-a68e-d9137c3b25bf-kube-api-access-zldkj\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070405 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-config\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070419 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-dir\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070432 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070617 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dtjpz"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.070435 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-serving-cert\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.071041 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45305c05-bdb6-4b38-acdb-2fb7f472ed68-trusted-ca\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.071062 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.071080 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8b87c1a-1579-4415-8134-021c1d7c390e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.071108 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-encryption-config\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.071123 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wts7z\" (UniqueName: \"kubernetes.io/projected/9ef6e180-79df-40ee-bf2b-96332e35c052-kube-api-access-wts7z\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.071227 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.076325 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537016-rt648"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.094081 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.094281 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.099242 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.105147 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.105492 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-clhjs"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.105726 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.106012 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9kxx2"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.106029 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.106321 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-t69c7"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.106435 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.106455 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.107051 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gbl86"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.107077 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zsp52"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.107122 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.107189 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.108708 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.108893 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.109972 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.114429 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.120718 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.121299 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.121645 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.122343 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.122548 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mflth"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.122895 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.129779 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.130075 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.131998 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.132593 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.133430 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.147956 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.149080 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.164722 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.167014 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.169111 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171514 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qfgcq"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171836 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171864 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfgdm\" (UniqueName: \"kubernetes.io/projected/d18ed308-bc67-4689-8926-3ceb3b799252-kube-api-access-lfgdm\") pod \"dns-operator-744455d44c-fszhl\" (UID: \"d18ed308-bc67-4689-8926-3ceb3b799252\") " pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171882 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fjvn\" (UniqueName: \"kubernetes.io/projected/365ad46d-8d32-45b7-bcbe-e4b771babb9a-kube-api-access-9fjvn\") pod \"migrator-59844c95c7-w4ph9\" (UID: \"365ad46d-8d32-45b7-bcbe-e4b771babb9a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171906 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171923 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/710bae1b-9785-4c51-a128-f91f4e962113-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171940 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4v9v\" (UniqueName: \"kubernetes.io/projected/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-kube-api-access-h4v9v\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171955 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45305c05-bdb6-4b38-acdb-2fb7f472ed68-bound-sa-token\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171971 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c828db27-a26e-41dd-a68e-d9137c3b25bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.171988 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-service-ca-bundle\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172004 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f71939-65e0-4459-9c58-211f5cafd259-serving-cert\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172019 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/444f893d-0c15-4a0c-8805-38f51f8037f6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-729pl\" (UID: \"444f893d-0c15-4a0c-8805-38f51f8037f6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172035 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-stats-auth\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172059 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172075 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e14522ed-1f77-462c-a5ec-caa1efe74a35-node-pullsecrets\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172090 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-client-ca\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172106 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172120 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/710bae1b-9785-4c51-a128-f91f4e962113-serving-cert\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172135 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-config\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172152 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwwsg\" (UniqueName: \"kubernetes.io/projected/f8b87c1a-1579-4415-8134-021c1d7c390e-kube-api-access-xwwsg\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172166 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk262\" (UniqueName: \"kubernetes.io/projected/444f893d-0c15-4a0c-8805-38f51f8037f6-kube-api-access-gk262\") pod \"cluster-samples-operator-665b6dd947-729pl\" (UID: \"444f893d-0c15-4a0c-8805-38f51f8037f6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172181 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172197 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9jm2\" (UniqueName: \"kubernetes.io/projected/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-kube-api-access-s9jm2\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172211 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d18ed308-bc67-4689-8926-3ceb3b799252-metrics-tls\") pod \"dns-operator-744455d44c-fszhl\" (UID: \"d18ed308-bc67-4689-8926-3ceb3b799252\") " pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172228 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zldkj\" (UniqueName: \"kubernetes.io/projected/c828db27-a26e-41dd-a68e-d9137c3b25bf-kube-api-access-zldkj\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172244 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-config\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172257 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-dir\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172272 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/202949b2-ad8d-451d-8ca3-0f94de759612-config\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172288 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-serving-cert\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172303 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45305c05-bdb6-4b38-acdb-2fb7f472ed68-trusted-ca\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172317 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172334 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8b87c1a-1579-4415-8134-021c1d7c390e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172350 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2gn8\" (UniqueName: \"kubernetes.io/projected/d51493b0-bd22-4a0e-9f54-c8a980289a4d-kube-api-access-r2gn8\") pod \"control-plane-machine-set-operator-78cbb6b69f-kncvh\" (UID: \"d51493b0-bd22-4a0e-9f54-c8a980289a4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172372 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-encryption-config\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172388 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wts7z\" (UniqueName: \"kubernetes.io/projected/9ef6e180-79df-40ee-bf2b-96332e35c052-kube-api-access-wts7z\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172404 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-serving-cert\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172418 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-trusted-ca\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172432 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-trusted-ca-bundle\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172446 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45305c05-bdb6-4b38-acdb-2fb7f472ed68-metrics-tls\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172459 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-config\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172479 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-etcd-client\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172493 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-etcd-serving-ca\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172508 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-config\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172523 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c828db27-a26e-41dd-a68e-d9137c3b25bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172537 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-auth-proxy-config\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172552 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d51493b0-bd22-4a0e-9f54-c8a980289a4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-kncvh\" (UID: \"d51493b0-bd22-4a0e-9f54-c8a980289a4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172569 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-metrics-certs\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172584 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5986cf5-a400-423c-916d-88260072d408-etcd-client\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172600 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m97gl\" (UniqueName: \"kubernetes.io/projected/b6f71939-65e0-4459-9c58-211f5cafd259-kube-api-access-m97gl\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172616 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172630 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172645 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-config\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172679 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-default-certificate\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172695 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-config\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172725 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jz5j\" (UniqueName: \"kubernetes.io/projected/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-kube-api-access-9jz5j\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172743 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-oauth-config\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172759 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-images\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172775 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qphj7\" (UniqueName: \"kubernetes.io/projected/d5986cf5-a400-423c-916d-88260072d408-kube-api-access-qphj7\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172810 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlmx5\" (UniqueName: \"kubernetes.io/projected/45305c05-bdb6-4b38-acdb-2fb7f472ed68-kube-api-access-vlmx5\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172826 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-etcd-ca\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172842 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172856 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-image-import-ca\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172871 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-serving-cert\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172887 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-service-ca-bundle\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172902 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/202949b2-ad8d-451d-8ca3-0f94de759612-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172918 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce8c6777-cb48-40db-b97c-e79398a0059e-serving-cert\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172934 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-audit-dir\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172949 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvzgj\" (UniqueName: \"kubernetes.io/projected/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-kube-api-access-hvzgj\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172965 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-machine-approver-tls\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172981 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcf6w\" (UniqueName: \"kubernetes.io/projected/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-kube-api-access-wcf6w\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.172996 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wksrn\" (UniqueName: \"kubernetes.io/projected/e105188e-f3d1-4757-8134-0db7ce5cbbf2-kube-api-access-wksrn\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173010 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-audit-policies\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173026 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173041 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-config\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173056 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-audit\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173072 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9tvh\" (UniqueName: \"kubernetes.io/projected/710bae1b-9785-4c51-a128-f91f4e962113-kube-api-access-d9tvh\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173087 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5986cf5-a400-423c-916d-88260072d408-serving-cert\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173103 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-oauth-serving-cert\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173119 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99hh7\" (UniqueName: \"kubernetes.io/projected/e14522ed-1f77-462c-a5ec-caa1efe74a35-kube-api-access-99hh7\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173140 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173158 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-config\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173174 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c376125-89b2-43d4-970a-c13ff4fbed59-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173189 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c376125-89b2-43d4-970a-c13ff4fbed59-config\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173205 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-etcd-service-ca\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173220 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-etcd-client\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173236 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-policies\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173252 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173273 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-config\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173290 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-service-ca\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173305 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173319 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8b87c1a-1579-4415-8134-021c1d7c390e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173336 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l5dd\" (UniqueName: \"kubernetes.io/projected/1c376125-89b2-43d4-970a-c13ff4fbed59-kube-api-access-6l5dd\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173350 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-serving-cert\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173365 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173380 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-client-ca\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173396 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ef6e180-79df-40ee-bf2b-96332e35c052-serving-cert\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173410 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-encryption-config\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173424 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e14522ed-1f77-462c-a5ec-caa1efe74a35-audit-dir\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173441 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zls89\" (UniqueName: \"kubernetes.io/projected/a4baae9f-93ba-40bf-9c28-01f13cdcaf43-kube-api-access-zls89\") pod \"downloads-7954f5f757-82v8f\" (UID: \"a4baae9f-93ba-40bf-9c28-01f13cdcaf43\") " pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173455 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8b87c1a-1579-4415-8134-021c1d7c390e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173470 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv96b\" (UniqueName: \"kubernetes.io/projected/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-kube-api-access-zv96b\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173486 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173501 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173515 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173532 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/202949b2-ad8d-451d-8ca3-0f94de759612-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.173547 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsfw6\" (UniqueName: \"kubernetes.io/projected/ce8c6777-cb48-40db-b97c-e79398a0059e-kube-api-access-wsfw6\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.177638 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v6dfv"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.178371 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.181520 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-trusted-ca-bundle\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.181709 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45305c05-bdb6-4b38-acdb-2fb7f472ed68-trusted-ca\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.183181 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.183229 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qmwh7"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.184011 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-etcd-serving-ca\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.184357 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.184414 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.184723 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-config\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.185006 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-etcd-client\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.185027 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-config\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.186408 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f71939-65e0-4459-9c58-211f5cafd259-serving-cert\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.186930 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-oauth-config\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.187897 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c376125-89b2-43d4-970a-c13ff4fbed59-config\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.188379 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-config\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.188416 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/710bae1b-9785-4c51-a128-f91f4e962113-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.188525 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8b87c1a-1579-4415-8134-021c1d7c390e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.188704 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.189237 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.189274 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-serving-cert\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.189586 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-config\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.189628 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.190393 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e14522ed-1f77-462c-a5ec-caa1efe74a35-node-pullsecrets\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.190391 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.190443 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vgnmn"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.190591 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-trusted-ca\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.190816 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-audit-dir\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.190934 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-dir\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191188 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-policies\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191243 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45305c05-bdb6-4b38-acdb-2fb7f472ed68-metrics-tls\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191246 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-audit-policies\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191567 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-images\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191812 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pgbmd"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191970 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c828db27-a26e-41dd-a68e-d9137c3b25bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.191990 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.192038 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e14522ed-1f77-462c-a5ec-caa1efe74a35-audit-dir\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.192038 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c376125-89b2-43d4-970a-c13ff4fbed59-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.192393 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-auth-proxy-config\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.192618 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-config\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.192738 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-config\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.193399 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-client-ca\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.193472 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.193609 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.193697 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-config\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.193936 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/444f893d-0c15-4a0c-8805-38f51f8037f6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-729pl\" (UID: \"444f893d-0c15-4a0c-8805-38f51f8037f6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.194031 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-oauth-serving-cert\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.194361 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e105188e-f3d1-4757-8134-0db7ce5cbbf2-service-ca\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.194553 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-image-import-ca\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.194878 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c828db27-a26e-41dd-a68e-d9137c3b25bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.194924 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.195133 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e14522ed-1f77-462c-a5ec-caa1efe74a35-audit\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.195361 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.195370 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.195589 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-encryption-config\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.195828 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-etcd-client\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.195833 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/710bae1b-9785-4c51-a128-f91f4e962113-serving-cert\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.196330 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-serving-cert\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.196629 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.196694 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbbmg"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.197584 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-client-ca\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.198347 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.199331 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.202067 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.202521 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-encryption-config\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203005 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e105188e-f3d1-4757-8134-0db7ce5cbbf2-console-serving-cert\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203050 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203055 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203297 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203363 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8b87c1a-1579-4415-8134-021c1d7c390e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203581 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e14522ed-1f77-462c-a5ec-caa1efe74a35-serving-cert\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203587 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-machine-approver-tls\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.203936 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ef6e180-79df-40ee-bf2b-96332e35c052-serving-cert\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.204941 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6rfkk"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.205972 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.206802 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-txbn6"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.207926 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.207959 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fszhl"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.214789 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.215149 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf6mj"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.216621 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dtjpz"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.217601 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.218807 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537016-rt648"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.219636 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.220045 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mflth"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.221083 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.222559 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-82v8f"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.226015 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.227569 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.227745 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.229766 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pc7k8"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.230935 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.232569 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.233879 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.235733 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.237178 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.239027 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.240506 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-clhjs"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.241712 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.242953 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zsp52"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.244378 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.254774 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-t69c7"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.257012 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.257555 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-txbn6"] Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.268582 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279392 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5986cf5-a400-423c-916d-88260072d408-serving-cert\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279434 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279455 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-etcd-service-ca\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279484 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/202949b2-ad8d-451d-8ca3-0f94de759612-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279499 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsfw6\" (UniqueName: \"kubernetes.io/projected/ce8c6777-cb48-40db-b97c-e79398a0059e-kube-api-access-wsfw6\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279529 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfgdm\" (UniqueName: \"kubernetes.io/projected/d18ed308-bc67-4689-8926-3ceb3b799252-kube-api-access-lfgdm\") pod \"dns-operator-744455d44c-fszhl\" (UID: \"d18ed308-bc67-4689-8926-3ceb3b799252\") " pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279547 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fjvn\" (UniqueName: \"kubernetes.io/projected/365ad46d-8d32-45b7-bcbe-e4b771babb9a-kube-api-access-9fjvn\") pod \"migrator-59844c95c7-w4ph9\" (UID: \"365ad46d-8d32-45b7-bcbe-e4b771babb9a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279584 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-service-ca-bundle\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279599 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-stats-auth\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279617 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-config\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279644 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9jm2\" (UniqueName: \"kubernetes.io/projected/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-kube-api-access-s9jm2\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279682 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d18ed308-bc67-4689-8926-3ceb3b799252-metrics-tls\") pod \"dns-operator-744455d44c-fszhl\" (UID: \"d18ed308-bc67-4689-8926-3ceb3b799252\") " pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279704 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/202949b2-ad8d-451d-8ca3-0f94de759612-config\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279722 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2gn8\" (UniqueName: \"kubernetes.io/projected/d51493b0-bd22-4a0e-9f54-c8a980289a4d-kube-api-access-r2gn8\") pod \"control-plane-machine-set-operator-78cbb6b69f-kncvh\" (UID: \"d51493b0-bd22-4a0e-9f54-c8a980289a4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279757 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d51493b0-bd22-4a0e-9f54-c8a980289a4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-kncvh\" (UID: \"d51493b0-bd22-4a0e-9f54-c8a980289a4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279774 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5986cf5-a400-423c-916d-88260072d408-etcd-client\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279790 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-metrics-certs\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279804 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-default-certificate\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279818 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-config\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279855 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qphj7\" (UniqueName: \"kubernetes.io/projected/d5986cf5-a400-423c-916d-88260072d408-kube-api-access-qphj7\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279886 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-etcd-ca\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279902 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-service-ca-bundle\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.279919 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/202949b2-ad8d-451d-8ca3-0f94de759612-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.280107 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce8c6777-cb48-40db-b97c-e79398a0059e-serving-cert\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.280195 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-service-ca-bundle\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.280331 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/202949b2-ad8d-451d-8ca3-0f94de759612-config\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.280358 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.280737 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce8c6777-cb48-40db-b97c-e79398a0059e-config\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.282736 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce8c6777-cb48-40db-b97c-e79398a0059e-serving-cert\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.288861 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.291573 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-config\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.292495 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/202949b2-ad8d-451d-8ca3-0f94de759612-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.308092 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.310137 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-etcd-service-ca\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.328288 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.331504 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d5986cf5-a400-423c-916d-88260072d408-etcd-ca\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.348448 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.354672 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5986cf5-a400-423c-916d-88260072d408-serving-cert\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.368120 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.388155 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.408096 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.413022 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5986cf5-a400-423c-916d-88260072d408-etcd-client\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.428087 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.450078 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.467775 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.472610 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d18ed308-bc67-4689-8926-3ceb3b799252-metrics-tls\") pod \"dns-operator-744455d44c-fszhl\" (UID: \"d18ed308-bc67-4689-8926-3ceb3b799252\") " pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.488466 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.507798 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.528326 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.568157 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.588580 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.608228 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.628169 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.648351 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.655356 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-metrics-certs\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.669026 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.674112 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-stats-auth\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.688569 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.691435 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-service-ca-bundle\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.708786 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.728183 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.735239 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-default-certificate\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.749168 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.768045 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.789621 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.809404 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.829462 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.848797 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.869770 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.873829 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d51493b0-bd22-4a0e-9f54-c8a980289a4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-kncvh\" (UID: \"d51493b0-bd22-4a0e-9f54-c8a980289a4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.888184 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.908176 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.927863 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.949065 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.968714 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 19:37:23 crc kubenswrapper[4839]: I0227 19:37:23.989053 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.007953 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.029369 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.048332 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.068438 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.107132 4839 request.go:700] Waited for 1.001077172s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.108841 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.128382 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.153860 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.168710 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.190250 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.208448 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.229254 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.249262 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.269387 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.288838 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.308192 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.327896 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.348266 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.367740 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.389281 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.408952 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.428799 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.448531 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.468123 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.488617 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.508894 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.528494 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.548385 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.568293 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.588491 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.609024 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.629859 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.648024 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.668450 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.688580 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.709042 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.729184 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.749538 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.769855 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.789388 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.808900 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.828207 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.849076 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.889761 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wts7z\" (UniqueName: \"kubernetes.io/projected/9ef6e180-79df-40ee-bf2b-96332e35c052-kube-api-access-wts7z\") pod \"controller-manager-879f6c89f-hbbmg\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.911107 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jz5j\" (UniqueName: \"kubernetes.io/projected/1fcaf0d6-2daf-4a11-b871-b718dc7f2541-kube-api-access-9jz5j\") pod \"console-operator-58897d9998-9kxx2\" (UID: \"1fcaf0d6-2daf-4a11-b871-b718dc7f2541\") " pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.922762 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m97gl\" (UniqueName: \"kubernetes.io/projected/b6f71939-65e0-4459-9c58-211f5cafd259-kube-api-access-m97gl\") pod \"route-controller-manager-6576b87f9c-gm67d\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.945929 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwwsg\" (UniqueName: \"kubernetes.io/projected/f8b87c1a-1579-4415-8134-021c1d7c390e-kube-api-access-xwwsg\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.966554 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk262\" (UniqueName: \"kubernetes.io/projected/444f893d-0c15-4a0c-8805-38f51f8037f6-kube-api-access-gk262\") pod \"cluster-samples-operator-665b6dd947-729pl\" (UID: \"444f893d-0c15-4a0c-8805-38f51f8037f6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:24 crc kubenswrapper[4839]: I0227 19:37:24.986070 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99hh7\" (UniqueName: \"kubernetes.io/projected/e14522ed-1f77-462c-a5ec-caa1efe74a35-kube-api-access-99hh7\") pod \"apiserver-76f77b778f-qfgcq\" (UID: \"e14522ed-1f77-462c-a5ec-caa1efe74a35\") " pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.002209 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45305c05-bdb6-4b38-acdb-2fb7f472ed68-bound-sa-token\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.018492 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.031111 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4v9v\" (UniqueName: \"kubernetes.io/projected/e7326984-1e3a-40e6-920d-ffdd7a2aaa89-kube-api-access-h4v9v\") pod \"apiserver-7bbb656c7d-xm448\" (UID: \"e7326984-1e3a-40e6-920d-ffdd7a2aaa89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.039409 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.043161 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zldkj\" (UniqueName: \"kubernetes.io/projected/c828db27-a26e-41dd-a68e-d9137c3b25bf-kube-api-access-zldkj\") pod \"openshift-controller-manager-operator-756b6f6bc6-bvwws\" (UID: \"c828db27-a26e-41dd-a68e-d9137c3b25bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.063351 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlmx5\" (UniqueName: \"kubernetes.io/projected/45305c05-bdb6-4b38-acdb-2fb7f472ed68-kube-api-access-vlmx5\") pod \"ingress-operator-5b745b69d9-27r5h\" (UID: \"45305c05-bdb6-4b38-acdb-2fb7f472ed68\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.075931 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.083297 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcf6w\" (UniqueName: \"kubernetes.io/projected/f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad-kube-api-access-wcf6w\") pod \"machine-approver-56656f9798-xf9jb\" (UID: \"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.102727 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wksrn\" (UniqueName: \"kubernetes.io/projected/e105188e-f3d1-4757-8134-0db7ce5cbbf2-kube-api-access-wksrn\") pod \"console-f9d7485db-pc7k8\" (UID: \"e105188e-f3d1-4757-8134-0db7ce5cbbf2\") " pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.126598 4839 request.go:700] Waited for 1.935618184s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-config-operator/serviceaccounts/openshift-config-operator/token Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.131643 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvzgj\" (UniqueName: \"kubernetes.io/projected/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-kube-api-access-hvzgj\") pod \"oauth-openshift-558db77b4-v6dfv\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.145417 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9tvh\" (UniqueName: \"kubernetes.io/projected/710bae1b-9785-4c51-a128-f91f4e962113-kube-api-access-d9tvh\") pod \"openshift-config-operator-7777fb866f-gbl86\" (UID: \"710bae1b-9785-4c51-a128-f91f4e962113\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.159888 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.170185 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8b87c1a-1579-4415-8134-021c1d7c390e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vk2tn\" (UID: \"f8b87c1a-1579-4415-8134-021c1d7c390e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.173603 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.179215 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.184435 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zls89\" (UniqueName: \"kubernetes.io/projected/a4baae9f-93ba-40bf-9c28-01f13cdcaf43-kube-api-access-zls89\") pod \"downloads-7954f5f757-82v8f\" (UID: \"a4baae9f-93ba-40bf-9c28-01f13cdcaf43\") " pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.204462 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.206922 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9c869a1_0bf6_49a4_a0f9_12f7f9b637ad.slice/crio-9b1249d17fc201119cd5c31d3df8efc9ed60b4a3362271dbd3c71127e21cdefb WatchSource:0}: Error finding container 9b1249d17fc201119cd5c31d3df8efc9ed60b4a3362271dbd3c71127e21cdefb: Status 404 returned error can't find the container with id 9b1249d17fc201119cd5c31d3df8efc9ed60b4a3362271dbd3c71127e21cdefb Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.208799 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv96b\" (UniqueName: \"kubernetes.io/projected/22e4cd58-dab8-41fa-9cae-323ce54a4b0d-kube-api-access-zv96b\") pod \"machine-api-operator-5694c8668f-pgbmd\" (UID: \"22e4cd58-dab8-41fa-9cae-323ce54a4b0d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.215524 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.223193 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l5dd\" (UniqueName: \"kubernetes.io/projected/1c376125-89b2-43d4-970a-c13ff4fbed59-kube-api-access-6l5dd\") pod \"openshift-apiserver-operator-796bbdcf4f-67h5d\" (UID: \"1c376125-89b2-43d4-970a-c13ff4fbed59\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.227342 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.230092 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.240714 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.247074 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9kxx2"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.248648 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.254086 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.269034 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.269127 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.275929 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.292028 4839 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.304576 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbbmg"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.309174 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.310900 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.327239 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.331779 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.344215 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" event={"ID":"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad","Type":"ContainerStarted","Data":"9b1249d17fc201119cd5c31d3df8efc9ed60b4a3362271dbd3c71127e21cdefb"} Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.354436 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" event={"ID":"1fcaf0d6-2daf-4a11-b871-b718dc7f2541","Type":"ContainerStarted","Data":"029797dc62565e6ac2cf530ee6614a70b4e0ee0fcd7ebc09aec1cbd64082c04a"} Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.361770 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/202949b2-ad8d-451d-8ca3-0f94de759612-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qlhjn\" (UID: \"202949b2-ad8d-451d-8ca3-0f94de759612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.389358 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsfw6\" (UniqueName: \"kubernetes.io/projected/ce8c6777-cb48-40db-b97c-e79398a0059e-kube-api-access-wsfw6\") pod \"authentication-operator-69f744f599-vgnmn\" (UID: \"ce8c6777-cb48-40db-b97c-e79398a0059e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.392326 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ef6e180_79df_40ee_bf2b_96332e35c052.slice/crio-79446c82f49ca00e64f617e1bf04a1499f6134e9fa86aaa22d9e46576f2196b4 WatchSource:0}: Error finding container 79446c82f49ca00e64f617e1bf04a1499f6134e9fa86aaa22d9e46576f2196b4: Status 404 returned error can't find the container with id 79446c82f49ca00e64f617e1bf04a1499f6134e9fa86aaa22d9e46576f2196b4 Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.394479 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.399534 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6f71939_65e0_4459_9c58_211f5cafd259.slice/crio-5b92219f064e325beab6765f8e32961736356c11e77b621078b21bc7cb4e06c6 WatchSource:0}: Error finding container 5b92219f064e325beab6765f8e32961736356c11e77b621078b21bc7cb4e06c6: Status 404 returned error can't find the container with id 5b92219f064e325beab6765f8e32961736356c11e77b621078b21bc7cb4e06c6 Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.414057 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fjvn\" (UniqueName: \"kubernetes.io/projected/365ad46d-8d32-45b7-bcbe-e4b771babb9a-kube-api-access-9fjvn\") pod \"migrator-59844c95c7-w4ph9\" (UID: \"365ad46d-8d32-45b7-bcbe-e4b771babb9a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.425238 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.432161 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfgdm\" (UniqueName: \"kubernetes.io/projected/d18ed308-bc67-4689-8926-3ceb3b799252-kube-api-access-lfgdm\") pod \"dns-operator-744455d44c-fszhl\" (UID: \"d18ed308-bc67-4689-8926-3ceb3b799252\") " pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.443347 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9jm2\" (UniqueName: \"kubernetes.io/projected/81d8a2c8-0ce3-4a9e-8d41-370b2107a840-kube-api-access-s9jm2\") pod \"router-default-5444994796-85jfr\" (UID: \"81d8a2c8-0ce3-4a9e-8d41-370b2107a840\") " pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.456979 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gbl86"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.473204 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2gn8\" (UniqueName: \"kubernetes.io/projected/d51493b0-bd22-4a0e-9f54-c8a980289a4d-kube-api-access-r2gn8\") pod \"control-plane-machine-set-operator-78cbb6b69f-kncvh\" (UID: \"d51493b0-bd22-4a0e-9f54-c8a980289a4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.491978 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qphj7\" (UniqueName: \"kubernetes.io/projected/d5986cf5-a400-423c-916d-88260072d408-kube-api-access-qphj7\") pod \"etcd-operator-b45778765-qmwh7\" (UID: \"d5986cf5-a400-423c-916d-88260072d408\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.508363 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod710bae1b_9785_4c51_a128_f91f4e962113.slice/crio-288116ee90c2e687af8200f36bc1385a821621bb8125fed9ba0fbc0c9bc2a818 WatchSource:0}: Error finding container 288116ee90c2e687af8200f36bc1385a821621bb8125fed9ba0fbc0c9bc2a818: Status 404 returned error can't find the container with id 288116ee90c2e687af8200f36bc1385a821621bb8125fed9ba0fbc0c9bc2a818 Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.511447 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.516475 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.609873 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npz26\" (UniqueName: \"kubernetes.io/projected/c1203271-c1c1-42e0-ab24-7403cc85726d-kube-api-access-npz26\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610220 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-images\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610293 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610322 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chttk\" (UniqueName: \"kubernetes.io/projected/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-kube-api-access-chttk\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610343 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dtjpz\" (UID: \"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610368 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1203271-c1c1-42e0-ab24-7403cc85726d-proxy-tls\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610761 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/df78f97a-5f5a-400e-a898-8c656da2c72f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dk4jt\" (UID: \"df78f97a-5f5a-400e-a898-8c656da2c72f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610796 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610811 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp2zk\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-kube-api-access-cp2zk\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610837 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610853 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610872 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1203271-c1c1-42e0-ab24-7403cc85726d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610941 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74wdp\" (UniqueName: \"kubernetes.io/projected/df78f97a-5f5a-400e-a898-8c656da2c72f-kube-api-access-74wdp\") pod \"package-server-manager-789f6589d5-dk4jt\" (UID: \"df78f97a-5f5a-400e-a898-8c656da2c72f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610974 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-tls\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.610989 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghcsb\" (UniqueName: \"kubernetes.io/projected/c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f-kube-api-access-ghcsb\") pod \"multus-admission-controller-857f4d67dd-dtjpz\" (UID: \"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611004 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-proxy-tls\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611018 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx6vv\" (UniqueName: \"kubernetes.io/projected/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-kube-api-access-gx6vv\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611044 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-certificates\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611063 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611080 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-trusted-ca\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611094 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-bound-sa-token\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.611123 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: E0227 19:37:25.612787 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.112766419 +0000 UTC m=+207.757636244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.642423 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.648111 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.654911 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.665556 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.674937 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.710160 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81d8a2c8_0ce3_4a9e_8d41_370b2107a840.slice/crio-8d3829e3c9d2137c25eb569ed4f2760784455f788ce923c0ee9b90b574169307 WatchSource:0}: Error finding container 8d3829e3c9d2137c25eb569ed4f2760784455f788ce923c0ee9b90b574169307: Status 404 returned error can't find the container with id 8d3829e3c9d2137c25eb569ed4f2760784455f788ce923c0ee9b90b574169307 Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.710182 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.712603 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:25 crc kubenswrapper[4839]: E0227 19:37:25.712739 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.212716955 +0000 UTC m=+207.857586700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.712940 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.713004 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/93a37e84-0fbe-4825-a605-968b25a9eca2-signing-key\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.713625 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/22eb69e2-54a0-4214-825e-0d3c5d3da62a-node-bootstrap-token\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.714299 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npz26\" (UniqueName: \"kubernetes.io/projected/c1203271-c1c1-42e0-ab24-7403cc85726d-kube-api-access-npz26\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.714363 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx27s\" (UniqueName: \"kubernetes.io/projected/4a9a60f2-147d-49c1-beaa-7f13ad42737a-kube-api-access-wx27s\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.714432 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.714553 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-metrics-tls\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715252 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715278 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-csi-data-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715310 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715620 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b927691-5423-4670-8443-c0a6a9d3bc13-config-volume\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715649 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-config\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715736 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6ddc4982-fe8a-4121-9dc3-b71600c77843-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715938 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-images\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715973 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft24q\" (UniqueName: \"kubernetes.io/projected/22eb69e2-54a0-4214-825e-0d3c5d3da62a-kube-api-access-ft24q\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.715997 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-socket-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716039 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716061 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00d587ff-502e-42ed-9d00-02f639086a6d-cert\") pod \"ingress-canary-zsp52\" (UID: \"00d587ff-502e-42ed-9d00-02f639086a6d\") " pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716096 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4a9a60f2-147d-49c1-beaa-7f13ad42737a-tmpfs\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716115 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-plugins-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716152 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jdxz\" (UniqueName: \"kubernetes.io/projected/4b927691-5423-4670-8443-c0a6a9d3bc13-kube-api-access-2jdxz\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716190 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b927691-5423-4670-8443-c0a6a9d3bc13-secret-volume\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716232 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chttk\" (UniqueName: \"kubernetes.io/projected/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-kube-api-access-chttk\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716257 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dtjpz\" (UID: \"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:25 crc kubenswrapper[4839]: E0227 19:37:25.716428 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.216413045 +0000 UTC m=+207.861282850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716505 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1203271-c1c1-42e0-ab24-7403cc85726d-proxy-tls\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716768 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/df78f97a-5f5a-400e-a898-8c656da2c72f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dk4jt\" (UID: \"df78f97a-5f5a-400e-a898-8c656da2c72f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.716802 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/22eb69e2-54a0-4214-825e-0d3c5d3da62a-certs\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.717030 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a9a60f2-147d-49c1-beaa-7f13ad42737a-apiservice-cert\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.717101 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.717120 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp2zk\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-kube-api-access-cp2zk\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.717777 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.718332 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-images\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.718418 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.718526 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-mountpoint-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.718873 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.718921 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw2jc\" (UniqueName: \"kubernetes.io/projected/6ddc4982-fe8a-4121-9dc3-b71600c77843-kube-api-access-kw2jc\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.719788 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2zc9\" (UniqueName: \"kubernetes.io/projected/d8a6578e-ff7c-4cb6-a051-2dabed753669-kube-api-access-c2zc9\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.719824 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1203271-c1c1-42e0-ab24-7403cc85726d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.719847 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-srv-cert\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.721047 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/df78f97a-5f5a-400e-a898-8c656da2c72f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dk4jt\" (UID: \"df78f97a-5f5a-400e-a898-8c656da2c72f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.721237 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjwhv\" (UniqueName: \"kubernetes.io/projected/93a37e84-0fbe-4825-a605-968b25a9eca2-kube-api-access-vjwhv\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.721316 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.721580 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.722440 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1203271-c1c1-42e0-ab24-7403cc85726d-proxy-tls\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.723134 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8d85\" (UniqueName: \"kubernetes.io/projected/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-kube-api-access-k8d85\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.723186 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-config-volume\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.723251 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/93a37e84-0fbe-4825-a605-968b25a9eca2-signing-cabundle\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.723277 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmg87\" (UniqueName: \"kubernetes.io/projected/95e89fb2-3271-46bd-878b-6ff642b17214-kube-api-access-kmg87\") pod \"auto-csr-approver-29537016-rt648\" (UID: \"95e89fb2-3271-46bd-878b-6ff642b17214\") " pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.725035 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74wdp\" (UniqueName: \"kubernetes.io/projected/df78f97a-5f5a-400e-a898-8c656da2c72f-kube-api-access-74wdp\") pod \"package-server-manager-789f6589d5-dk4jt\" (UID: \"df78f97a-5f5a-400e-a898-8c656da2c72f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.725106 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-profile-collector-cert\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.725131 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8a6578e-ff7c-4cb6-a051-2dabed753669-serving-cert\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.725535 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1203271-c1c1-42e0-ab24-7403cc85726d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.725999 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-registration-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726060 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-tls\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726108 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghcsb\" (UniqueName: \"kubernetes.io/projected/c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f-kube-api-access-ghcsb\") pod \"multus-admission-controller-857f4d67dd-dtjpz\" (UID: \"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726168 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-proxy-tls\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726192 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx6vv\" (UniqueName: \"kubernetes.io/projected/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-kube-api-access-gx6vv\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726419 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a9a60f2-147d-49c1-beaa-7f13ad42737a-webhook-cert\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726450 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-certificates\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726551 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89jxk\" (UniqueName: \"kubernetes.io/projected/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-kube-api-access-89jxk\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.726610 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727153 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727189 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk8m9\" (UniqueName: \"kubernetes.io/projected/00d587ff-502e-42ed-9d00-02f639086a6d-kube-api-access-rk8m9\") pod \"ingress-canary-zsp52\" (UID: \"00d587ff-502e-42ed-9d00-02f639086a6d\") " pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727266 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727300 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-trusted-ca\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727335 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6ddc4982-fe8a-4121-9dc3-b71600c77843-srv-cert\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727356 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-bound-sa-token\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.727981 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.728031 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8a6578e-ff7c-4cb6-a051-2dabed753669-config\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.728158 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.728187 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbkm8\" (UniqueName: \"kubernetes.io/projected/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-kube-api-access-dbkm8\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.728779 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqktt\" (UniqueName: \"kubernetes.io/projected/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-kube-api-access-pqktt\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.728950 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.729766 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-trusted-ca\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.730535 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.731583 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-proxy-tls\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.731645 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-tls\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.733727 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-certificates\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.735277 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dtjpz\" (UID: \"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.757081 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qfgcq"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.761037 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.768587 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v6dfv"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.775178 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npz26\" (UniqueName: \"kubernetes.io/projected/c1203271-c1c1-42e0-ab24-7403cc85726d-kube-api-access-npz26\") pod \"machine-config-controller-84d6567774-gxn9q\" (UID: \"c1203271-c1c1-42e0-ab24-7403cc85726d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.785150 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chttk\" (UniqueName: \"kubernetes.io/projected/c34af3d9-4299-46e1-ba6d-56210fd9e5e2-kube-api-access-chttk\") pod \"kube-storage-version-migrator-operator-b67b599dd-8qr2g\" (UID: \"c34af3d9-4299-46e1-ba6d-56210fd9e5e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.806492 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp2zk\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-kube-api-access-cp2zk\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.806923 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.825551 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74wdp\" (UniqueName: \"kubernetes.io/projected/df78f97a-5f5a-400e-a898-8c656da2c72f-kube-api-access-74wdp\") pod \"package-server-manager-789f6589d5-dk4jt\" (UID: \"df78f97a-5f5a-400e-a898-8c656da2c72f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.831551 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:25 crc kubenswrapper[4839]: E0227 19:37:25.832781 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.332757269 +0000 UTC m=+207.977627004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.833002 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode14522ed_1f77_462c_a5ec_caa1efe74a35.slice/crio-3891f3d3070abc77c35f42786265f2b30256652c33625a33e6ff12e4051fa0c4 WatchSource:0}: Error finding container 3891f3d3070abc77c35f42786265f2b30256652c33625a33e6ff12e4051fa0c4: Status 404 returned error can't find the container with id 3891f3d3070abc77c35f42786265f2b30256652c33625a33e6ff12e4051fa0c4 Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.835317 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45305c05_bdb6_4b38_acdb_2fb7f472ed68.slice/crio-6a2db90bd12c209b02841f43fde74a74adab71992c335d99c71b9e34191c79cb WatchSource:0}: Error finding container 6a2db90bd12c209b02841f43fde74a74adab71992c335d99c71b9e34191c79cb: Status 404 returned error can't find the container with id 6a2db90bd12c209b02841f43fde74a74adab71992c335d99c71b9e34191c79cb Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836125 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836200 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8d85\" (UniqueName: \"kubernetes.io/projected/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-kube-api-access-k8d85\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836232 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-config-volume\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836293 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/93a37e84-0fbe-4825-a605-968b25a9eca2-signing-cabundle\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836320 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmg87\" (UniqueName: \"kubernetes.io/projected/95e89fb2-3271-46bd-878b-6ff642b17214-kube-api-access-kmg87\") pod \"auto-csr-approver-29537016-rt648\" (UID: \"95e89fb2-3271-46bd-878b-6ff642b17214\") " pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836404 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-profile-collector-cert\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836465 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8a6578e-ff7c-4cb6-a051-2dabed753669-serving-cert\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836503 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-registration-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836577 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a9a60f2-147d-49c1-beaa-7f13ad42737a-webhook-cert\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836637 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89jxk\" (UniqueName: \"kubernetes.io/projected/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-kube-api-access-89jxk\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836695 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836699 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836793 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk8m9\" (UniqueName: \"kubernetes.io/projected/00d587ff-502e-42ed-9d00-02f639086a6d-kube-api-access-rk8m9\") pod \"ingress-canary-zsp52\" (UID: \"00d587ff-502e-42ed-9d00-02f639086a6d\") " pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836812 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836828 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6ddc4982-fe8a-4121-9dc3-b71600c77843-srv-cert\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836854 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8a6578e-ff7c-4cb6-a051-2dabed753669-config\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836879 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbkm8\" (UniqueName: \"kubernetes.io/projected/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-kube-api-access-dbkm8\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836898 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqktt\" (UniqueName: \"kubernetes.io/projected/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-kube-api-access-pqktt\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836915 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836940 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/93a37e84-0fbe-4825-a605-968b25a9eca2-signing-key\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836961 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/22eb69e2-54a0-4214-825e-0d3c5d3da62a-node-bootstrap-token\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.836980 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx27s\" (UniqueName: \"kubernetes.io/projected/4a9a60f2-147d-49c1-beaa-7f13ad42737a-kube-api-access-wx27s\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837001 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837025 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-metrics-tls\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837041 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837058 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-csi-data-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837081 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837099 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b927691-5423-4670-8443-c0a6a9d3bc13-config-volume\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837113 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-config\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837143 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6ddc4982-fe8a-4121-9dc3-b71600c77843-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837165 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft24q\" (UniqueName: \"kubernetes.io/projected/22eb69e2-54a0-4214-825e-0d3c5d3da62a-kube-api-access-ft24q\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837180 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-socket-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837210 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837258 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00d587ff-502e-42ed-9d00-02f639086a6d-cert\") pod \"ingress-canary-zsp52\" (UID: \"00d587ff-502e-42ed-9d00-02f639086a6d\") " pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837277 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4a9a60f2-147d-49c1-beaa-7f13ad42737a-tmpfs\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837316 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-plugins-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837338 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jdxz\" (UniqueName: \"kubernetes.io/projected/4b927691-5423-4670-8443-c0a6a9d3bc13-kube-api-access-2jdxz\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837354 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b927691-5423-4670-8443-c0a6a9d3bc13-secret-volume\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837373 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/22eb69e2-54a0-4214-825e-0d3c5d3da62a-certs\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837394 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a9a60f2-147d-49c1-beaa-7f13ad42737a-apiservice-cert\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837418 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-mountpoint-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837434 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw2jc\" (UniqueName: \"kubernetes.io/projected/6ddc4982-fe8a-4121-9dc3-b71600c77843-kube-api-access-kw2jc\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837454 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2zc9\" (UniqueName: \"kubernetes.io/projected/d8a6578e-ff7c-4cb6-a051-2dabed753669-kube-api-access-c2zc9\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837470 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-srv-cert\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837487 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjwhv\" (UniqueName: \"kubernetes.io/projected/93a37e84-0fbe-4825-a605-968b25a9eca2-kube-api-access-vjwhv\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.838497 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/93a37e84-0fbe-4825-a605-968b25a9eca2-signing-cabundle\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.839418 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-registration-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.840197 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8a6578e-ff7c-4cb6-a051-2dabed753669-config\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.840460 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6ddc4982-fe8a-4121-9dc3-b71600c77843-srv-cert\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.840536 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.837417 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-config-volume\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.841774 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-csi-data-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.841874 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.842614 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b927691-5423-4670-8443-c0a6a9d3bc13-config-volume\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.842769 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-socket-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.843780 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-profile-collector-cert\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.844093 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-mountpoint-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: E0227 19:37:25.844123 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.344085516 +0000 UTC m=+207.988955251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.846623 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6ddc4982-fe8a-4121-9dc3-b71600c77843-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.846726 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.847113 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.847288 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-plugins-dir\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.847625 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a9a60f2-147d-49c1-beaa-7f13ad42737a-webhook-cert\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.847647 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/93a37e84-0fbe-4825-a605-968b25a9eca2-signing-key\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.847975 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8a6578e-ff7c-4cb6-a051-2dabed753669-serving-cert\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.849034 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/22eb69e2-54a0-4214-825e-0d3c5d3da62a-node-bootstrap-token\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.850292 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-config\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.850991 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx6vv\" (UniqueName: \"kubernetes.io/projected/a783a0e0-5fb8-4542-b61c-bdd54fc3a331-kube-api-access-gx6vv\") pod \"machine-config-operator-74547568cd-kfdf5\" (UID: \"a783a0e0-5fb8-4542-b61c-bdd54fc3a331\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.852271 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/22eb69e2-54a0-4214-825e-0d3c5d3da62a-certs\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.857542 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4a9a60f2-147d-49c1-beaa-7f13ad42737a-tmpfs\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.860233 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-metrics-tls\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.860427 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-srv-cert\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.860981 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00d587ff-502e-42ed-9d00-02f639086a6d-cert\") pod \"ingress-canary-zsp52\" (UID: \"00d587ff-502e-42ed-9d00-02f639086a6d\") " pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.861666 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a9a60f2-147d-49c1-beaa-7f13ad42737a-apiservice-cert\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.863253 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc828db27_a26e_41dd_a68e_d9137c3b25bf.slice/crio-dc4ecc0e158d25c2dc743b30cde8b03368e13428b1f9ee73472d4a8976c94b0e WatchSource:0}: Error finding container dc4ecc0e158d25c2dc743b30cde8b03368e13428b1f9ee73472d4a8976c94b0e: Status 404 returned error can't find the container with id dc4ecc0e158d25c2dc743b30cde8b03368e13428b1f9ee73472d4a8976c94b0e Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.865136 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b927691-5423-4670-8443-c0a6a9d3bc13-secret-volume\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.868340 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghcsb\" (UniqueName: \"kubernetes.io/projected/c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f-kube-api-access-ghcsb\") pod \"multus-admission-controller-857f4d67dd-dtjpz\" (UID: \"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.882208 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-bound-sa-token\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.913627 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.938729 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:25 crc kubenswrapper[4839]: E0227 19:37:25.939263 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.439243139 +0000 UTC m=+208.084112874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.947843 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.948853 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8d85\" (UniqueName: \"kubernetes.io/projected/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-kube-api-access-k8d85\") pod \"marketplace-operator-79b997595-clhjs\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.949334 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.951854 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-82v8f"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.957053 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.958887 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pc7k8"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.959051 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjwhv\" (UniqueName: \"kubernetes.io/projected/93a37e84-0fbe-4825-a605-968b25a9eca2-kube-api-access-vjwhv\") pod \"service-ca-9c57cc56f-mflth\" (UID: \"93a37e84-0fbe-4825-a605-968b25a9eca2\") " pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.971370 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.973917 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk8m9\" (UniqueName: \"kubernetes.io/projected/00d587ff-502e-42ed-9d00-02f639086a6d-kube-api-access-rk8m9\") pod \"ingress-canary-zsp52\" (UID: \"00d587ff-502e-42ed-9d00-02f639086a6d\") " pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.981876 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vgnmn"] Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.982143 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" Feb 27 19:37:25 crc kubenswrapper[4839]: W0227 19:37:25.984755 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8b87c1a_1579_4415_8134_021c1d7c390e.slice/crio-a69bcc700b6d3b72ef7bdd21bf248739f97b2b71e93c40eba23ab81448555c5e WatchSource:0}: Error finding container a69bcc700b6d3b72ef7bdd21bf248739f97b2b71e93c40eba23ab81448555c5e: Status 404 returned error can't find the container with id a69bcc700b6d3b72ef7bdd21bf248739f97b2b71e93c40eba23ab81448555c5e Feb 27 19:37:25 crc kubenswrapper[4839]: I0227 19:37:25.989114 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b54c7b4f-365c-4088-bb22-94f6dc65c1c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v2gv2\" (UID: \"b54c7b4f-365c-4088-bb22-94f6dc65c1c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.002130 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode105188e_f3d1_4757_8134_0db7ce5cbbf2.slice/crio-1f2e3b7956d60ac0d879315d595d2823c6c7b31cafe0d8721725e0e6dfa4820c WatchSource:0}: Error finding container 1f2e3b7956d60ac0d879315d595d2823c6c7b31cafe0d8721725e0e6dfa4820c: Status 404 returned error can't find the container with id 1f2e3b7956d60ac0d879315d595d2823c6c7b31cafe0d8721725e0e6dfa4820c Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.003746 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.005945 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft24q\" (UniqueName: \"kubernetes.io/projected/22eb69e2-54a0-4214-825e-0d3c5d3da62a-kube-api-access-ft24q\") pod \"machine-config-server-6rfkk\" (UID: \"22eb69e2-54a0-4214-825e-0d3c5d3da62a\") " pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.017769 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7326984_1e3a_40e6_920d_ffdd7a2aaa89.slice/crio-a21600fe4a08d16ac9366dc0ff1b1dd6a689fce4a6d9bd84a08e215c6dba168b WatchSource:0}: Error finding container a21600fe4a08d16ac9366dc0ff1b1dd6a689fce4a6d9bd84a08e215c6dba168b: Status 404 returned error can't find the container with id a21600fe4a08d16ac9366dc0ff1b1dd6a689fce4a6d9bd84a08e215c6dba168b Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.019476 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qmwh7"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.021645 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.024003 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce8c6777_cb48_40db_b97c_e79398a0059e.slice/crio-f798f17d767b537b51f6cc460a70af8e9ba7dc6c93c82e613344bfb46cb64ac7 WatchSource:0}: Error finding container f798f17d767b537b51f6cc460a70af8e9ba7dc6c93c82e613344bfb46cb64ac7: Status 404 returned error can't find the container with id f798f17d767b537b51f6cc460a70af8e9ba7dc6c93c82e613344bfb46cb64ac7 Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.032027 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbkm8\" (UniqueName: \"kubernetes.io/projected/fb42fca6-045c-4c5d-8f0b-c29f40c746d0-kube-api-access-dbkm8\") pod \"catalog-operator-68c6474976-bs9xm\" (UID: \"fb42fca6-045c-4c5d-8f0b-c29f40c746d0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.034739 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.041953 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.042990 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.542973507 +0000 UTC m=+208.187843242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.048007 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.052947 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqktt\" (UniqueName: \"kubernetes.io/projected/aaca7144-ec64-4c3a-8fcd-b77b9d51fb89-kube-api-access-pqktt\") pod \"dns-default-t69c7\" (UID: \"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89\") " pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.063303 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pgbmd"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.065050 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.071049 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.077596 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmg87\" (UniqueName: \"kubernetes.io/projected/95e89fb2-3271-46bd-878b-6ff642b17214-kube-api-access-kmg87\") pod \"auto-csr-approver-29537016-rt648\" (UID: \"95e89fb2-3271-46bd-878b-6ff642b17214\") " pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.084783 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.093458 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89jxk\" (UniqueName: \"kubernetes.io/projected/1eb21c22-1c0f-4a33-93d8-d50cee6ecba4-kube-api-access-89jxk\") pod \"csi-hostpathplugin-txbn6\" (UID: \"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4\") " pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.093722 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.101423 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod365ad46d_8d32_45b7_bcbe_e4b771babb9a.slice/crio-8c3589734515a6ea49aa601748285af3fd64ca09add5af52d1130392c84e244d WatchSource:0}: Error finding container 8c3589734515a6ea49aa601748285af3fd64ca09add5af52d1130392c84e244d: Status 404 returned error can't find the container with id 8c3589734515a6ea49aa601748285af3fd64ca09add5af52d1130392c84e244d Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.101908 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zsp52" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.105915 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx27s\" (UniqueName: \"kubernetes.io/projected/4a9a60f2-147d-49c1-beaa-7f13ad42737a-kube-api-access-wx27s\") pod \"packageserver-d55dfcdfc-2cx5s\" (UID: \"4a9a60f2-147d-49c1-beaa-7f13ad42737a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.108656 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.129844 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mflth" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.137750 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.140405 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4x5z\" (UID: \"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.143411 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.143860 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.64384278 +0000 UTC m=+208.288712515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.149350 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2zc9\" (UniqueName: \"kubernetes.io/projected/d8a6578e-ff7c-4cb6-a051-2dabed753669-kube-api-access-c2zc9\") pod \"service-ca-operator-777779d784-s7t9g\" (UID: \"d8a6578e-ff7c-4cb6-a051-2dabed753669\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.154337 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6rfkk" Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.155944 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd51493b0_bd22_4a0e_9f54_c8a980289a4d.slice/crio-169a1ccac6749bc2ac9034833dc0627606f73a4410c7ad2ecce6e74454a5bca7 WatchSource:0}: Error finding container 169a1ccac6749bc2ac9034833dc0627606f73a4410c7ad2ecce6e74454a5bca7: Status 404 returned error can't find the container with id 169a1ccac6749bc2ac9034833dc0627606f73a4410c7ad2ecce6e74454a5bca7 Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.160658 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.169652 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw2jc\" (UniqueName: \"kubernetes.io/projected/6ddc4982-fe8a-4121-9dc3-b71600c77843-kube-api-access-kw2jc\") pod \"olm-operator-6b444d44fb-bpkff\" (UID: \"6ddc4982-fe8a-4121-9dc3-b71600c77843\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.190400 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jdxz\" (UniqueName: \"kubernetes.io/projected/4b927691-5423-4670-8443-c0a6a9d3bc13-kube-api-access-2jdxz\") pod \"collect-profiles-29537010-7cdv5\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.241274 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fszhl"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.244660 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.245024 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.745004902 +0000 UTC m=+208.389874637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.273121 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.329466 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.345263 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.345906 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.845884835 +0000 UTC m=+208.490754570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.354265 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.377135 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.382183 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-82v8f" event={"ID":"a4baae9f-93ba-40bf-9c28-01f13cdcaf43","Type":"ContainerStarted","Data":"61ab076a4a5135e7a6bad5aada236d052dcf0ed1cfd6329f603276ffb601d2db"} Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.389515 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd18ed308_bc67_4689_8926_3ceb3b799252.slice/crio-c55d0b3ef3dce1deaa6df76d43952c95558c3d394766c201f837bad69b12d9fc WatchSource:0}: Error finding container c55d0b3ef3dce1deaa6df76d43952c95558c3d394766c201f837bad69b12d9fc: Status 404 returned error can't find the container with id c55d0b3ef3dce1deaa6df76d43952c95558c3d394766c201f837bad69b12d9fc Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.394581 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" event={"ID":"f8b87c1a-1579-4415-8134-021c1d7c390e","Type":"ContainerStarted","Data":"a69bcc700b6d3b72ef7bdd21bf248739f97b2b71e93c40eba23ab81448555c5e"} Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.400474 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1203271_c1c1_42e0_ab24_7403cc85726d.slice/crio-61a1aa7804a7e80eb19fe8384325404a6c90135fc7fcd11a359b1566c6becc16 WatchSource:0}: Error finding container 61a1aa7804a7e80eb19fe8384325404a6c90135fc7fcd11a359b1566c6becc16: Status 404 returned error can't find the container with id 61a1aa7804a7e80eb19fe8384325404a6c90135fc7fcd11a359b1566c6becc16 Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.406982 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.408772 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" event={"ID":"b6f71939-65e0-4459-9c58-211f5cafd259","Type":"ContainerStarted","Data":"10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.408818 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" event={"ID":"b6f71939-65e0-4459-9c58-211f5cafd259","Type":"ContainerStarted","Data":"5b92219f064e325beab6765f8e32961736356c11e77b621078b21bc7cb4e06c6"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.409564 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.419913 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.422014 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.428048 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" event={"ID":"d5986cf5-a400-423c-916d-88260072d408","Type":"ContainerStarted","Data":"cff46423acfb9ead4e007c13e7cfc350635981e022f59a13915455defb04f9e2"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.431497 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" event={"ID":"1fcaf0d6-2daf-4a11-b871-b718dc7f2541","Type":"ContainerStarted","Data":"6639c1371d92860089f1fe489fb09b888ccb88996bd6141af19ed0200d84396b"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.432558 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.438045 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" event={"ID":"1c376125-89b2-43d4-970a-c13ff4fbed59","Type":"ContainerStarted","Data":"46ec93ec5aee8e52bf79d5478da66adfb11a69715061e856db587883229a6282"} Feb 27 19:37:26 crc kubenswrapper[4839]: W0227 19:37:26.438841 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda783a0e0_5fb8_4542_b61c_bdd54fc3a331.slice/crio-670e27a46a34cf3bffd694700526effa2b46c9913b128995b993cfb010a3889b WatchSource:0}: Error finding container 670e27a46a34cf3bffd694700526effa2b46c9913b128995b993cfb010a3889b: Status 404 returned error can't find the container with id 670e27a46a34cf3bffd694700526effa2b46c9913b128995b993cfb010a3889b Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.443429 4839 patch_prober.go:28] interesting pod/console-operator-58897d9998-9kxx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.443484 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" podUID="1fcaf0d6-2daf-4a11-b871-b718dc7f2541" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.443548 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" event={"ID":"c828db27-a26e-41dd-a68e-d9137c3b25bf","Type":"ContainerStarted","Data":"dc4ecc0e158d25c2dc743b30cde8b03368e13428b1f9ee73472d4a8976c94b0e"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.444969 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.448440 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.448726 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:26.948715746 +0000 UTC m=+208.593585481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.452049 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" event={"ID":"444f893d-0c15-4a0c-8805-38f51f8037f6","Type":"ContainerStarted","Data":"783a37292d2b40c66dfecd26db3be732ef45ed7bc0364707016d9dc515d31d0d"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.457339 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" event={"ID":"9ef6e180-79df-40ee-bf2b-96332e35c052","Type":"ContainerStarted","Data":"0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.457383 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" event={"ID":"9ef6e180-79df-40ee-bf2b-96332e35c052","Type":"ContainerStarted","Data":"79446c82f49ca00e64f617e1bf04a1499f6134e9fa86aaa22d9e46576f2196b4"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.457642 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.468895 4839 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-hbbmg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.468940 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" podUID="9ef6e180-79df-40ee-bf2b-96332e35c052" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.477969 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" event={"ID":"202949b2-ad8d-451d-8ca3-0f94de759612","Type":"ContainerStarted","Data":"cf2fc3a4db371a1e52b4a1ef99629ee633462d6be66b0e52369b11d6614d9008"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.479176 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" event={"ID":"365ad46d-8d32-45b7-bcbe-e4b771babb9a","Type":"ContainerStarted","Data":"8c3589734515a6ea49aa601748285af3fd64ca09add5af52d1130392c84e244d"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.492117 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" event={"ID":"e7326984-1e3a-40e6-920d-ffdd7a2aaa89","Type":"ContainerStarted","Data":"a21600fe4a08d16ac9366dc0ff1b1dd6a689fce4a6d9bd84a08e215c6dba168b"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.512615 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" event={"ID":"22e4cd58-dab8-41fa-9cae-323ce54a4b0d","Type":"ContainerStarted","Data":"43c3d00fade1784f5ef3055fb0ae8ba0386ab79c67f862432af758ab78be0c6b"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.526587 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" event={"ID":"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c","Type":"ContainerStarted","Data":"6df1b11c412c6bedd75d7a6c733af77084876f4f766f30e711db6c64659e8cd9"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.535008 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" event={"ID":"d51493b0-bd22-4a0e-9f54-c8a980289a4d","Type":"ContainerStarted","Data":"169a1ccac6749bc2ac9034833dc0627606f73a4410c7ad2ecce6e74454a5bca7"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.542648 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" event={"ID":"45305c05-bdb6-4b38-acdb-2fb7f472ed68","Type":"ContainerStarted","Data":"67b8e3a4b5a970d1e039bce1cffc4069efcf1de621c61c9de6553cd45933e0de"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.542716 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" event={"ID":"45305c05-bdb6-4b38-acdb-2fb7f472ed68","Type":"ContainerStarted","Data":"6a2db90bd12c209b02841f43fde74a74adab71992c335d99c71b9e34191c79cb"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.543496 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-clhjs"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.549038 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.550695 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.050661241 +0000 UTC m=+208.695530976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.570076 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-85jfr" event={"ID":"81d8a2c8-0ce3-4a9e-8d41-370b2107a840","Type":"ContainerStarted","Data":"1c15d0291067886388f4a4aecd98e02821959d6d17eb95c509e47a5171ad4b21"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.570119 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-85jfr" event={"ID":"81d8a2c8-0ce3-4a9e-8d41-370b2107a840","Type":"ContainerStarted","Data":"8d3829e3c9d2137c25eb569ed4f2760784455f788ce923c0ee9b90b574169307"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.572162 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" event={"ID":"ce8c6777-cb48-40db-b97c-e79398a0059e","Type":"ContainerStarted","Data":"f798f17d767b537b51f6cc460a70af8e9ba7dc6c93c82e613344bfb46cb64ac7"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.573890 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" event={"ID":"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad","Type":"ContainerStarted","Data":"23d13ba9af580d1a5d7b1a2629d263183b38a12a6dc1784fe26e6b9f0b0f399e"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.573916 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" event={"ID":"f9c869a1-0bf6-49a4-a0f9-12f7f9b637ad","Type":"ContainerStarted","Data":"dd09276d8b8d57b7abcb47635ff38e3d6672f0eb2e8ae23844da60fe86911f6d"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.579924 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" event={"ID":"e14522ed-1f77-462c-a5ec-caa1efe74a35","Type":"ContainerStarted","Data":"3891f3d3070abc77c35f42786265f2b30256652c33625a33e6ff12e4051fa0c4"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.582148 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pc7k8" event={"ID":"e105188e-f3d1-4757-8134-0db7ce5cbbf2","Type":"ContainerStarted","Data":"1f2e3b7956d60ac0d879315d595d2823c6c7b31cafe0d8721725e0e6dfa4820c"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.585099 4839 generic.go:334] "Generic (PLEG): container finished" podID="710bae1b-9785-4c51-a128-f91f4e962113" containerID="c0a1882114e84812038d7baf6d0e3d110c01fe2f3a1c97ce6d6653ac9c622e8a" exitCode=0 Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.585137 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" event={"ID":"710bae1b-9785-4c51-a128-f91f4e962113","Type":"ContainerDied","Data":"c0a1882114e84812038d7baf6d0e3d110c01fe2f3a1c97ce6d6653ac9c622e8a"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.585158 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" event={"ID":"710bae1b-9785-4c51-a128-f91f4e962113","Type":"ContainerStarted","Data":"288116ee90c2e687af8200f36bc1385a821621bb8125fed9ba0fbc0c9bc2a818"} Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.611183 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.650631 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.659179 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.159146931 +0000 UTC m=+208.804016666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.680032 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.683869 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:26 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:26 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:26 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.683905 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.752277 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.753348 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.253332385 +0000 UTC m=+208.898202120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.753484 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dtjpz"] Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.767403 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.853921 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.854265 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.354248638 +0000 UTC m=+208.999118373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:26 crc kubenswrapper[4839]: I0227 19:37:26.954625 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:26 crc kubenswrapper[4839]: E0227 19:37:26.955243 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.455227215 +0000 UTC m=+209.100096940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.047222 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm"] Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.053965 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2"] Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.056261 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.056572 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.556559001 +0000 UTC m=+209.201428726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.125884 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xf9jb" podStartSLOduration=164.125869585 podStartE2EDuration="2m44.125869585s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.124119593 +0000 UTC m=+208.768989328" watchObservedRunningTime="2026-02-27 19:37:27.125869585 +0000 UTC m=+208.770739320" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.145764 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mflth"] Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.157596 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.157998 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.657982591 +0000 UTC m=+209.302852326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.182167 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt"] Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.208953 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zsp52"] Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.210801 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-t69c7"] Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.267508 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.280722 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.780665963 +0000 UTC m=+209.425535698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.342189 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" podStartSLOduration=164.342176805 podStartE2EDuration="2m44.342176805s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.340267158 +0000 UTC m=+208.985136913" watchObservedRunningTime="2026-02-27 19:37:27.342176805 +0000 UTC m=+208.987046540" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.368654 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.369859 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.869809177 +0000 UTC m=+209.514678912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.370402 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.370754 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.870742535 +0000 UTC m=+209.515612270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.377129 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" podStartSLOduration=164.377114075 podStartE2EDuration="2m44.377114075s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.375077324 +0000 UTC m=+209.019947069" watchObservedRunningTime="2026-02-27 19:37:27.377114075 +0000 UTC m=+209.021983810" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.471902 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.472417 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:27.972398431 +0000 UTC m=+209.617268166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.596466 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.597359 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.097336021 +0000 UTC m=+209.742205756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.615273 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48560: no serving certificate available for the kubelet" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.639808 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" podStartSLOduration=164.639785985 podStartE2EDuration="2m44.639785985s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.637498606 +0000 UTC m=+209.282368361" watchObservedRunningTime="2026-02-27 19:37:27.639785985 +0000 UTC m=+209.284655730" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.682769 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48570: no serving certificate available for the kubelet" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.686316 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:27 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:27 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:27 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.686359 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.700339 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.700657 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.200640706 +0000 UTC m=+209.845510441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.785078 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48584: no serving certificate available for the kubelet" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.809546 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.817547 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.317527536 +0000 UTC m=+209.962397271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.824359 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pc7k8" event={"ID":"e105188e-f3d1-4757-8134-0db7ce5cbbf2","Type":"ContainerStarted","Data":"43830ecd5410ece9b1571ef39fc204dc0f4f210953088f9adce22ac402f11f54"} Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.879742 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zsp52" event={"ID":"00d587ff-502e-42ed-9d00-02f639086a6d","Type":"ContainerStarted","Data":"3efeb5212ff9fac81065d520e5aa0e59380ada79656d0a4fc127a73fadaadec3"} Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.892236 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" event={"ID":"c828db27-a26e-41dd-a68e-d9137c3b25bf","Type":"ContainerStarted","Data":"af37da2cc19409e79b1535014c56c81b1efd7ec839ee465faabcc7ffbce9bdae"} Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.910989 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.912023 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.411995999 +0000 UTC m=+210.056865724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.917462 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:27 crc kubenswrapper[4839]: E0227 19:37:27.920108 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.42009298 +0000 UTC m=+210.064962715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.924106 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48586: no serving certificate available for the kubelet" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.925215 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" event={"ID":"a783a0e0-5fb8-4542-b61c-bdd54fc3a331","Type":"ContainerStarted","Data":"670e27a46a34cf3bffd694700526effa2b46c9913b128995b993cfb010a3889b"} Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.933879 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" event={"ID":"d18ed308-bc67-4689-8926-3ceb3b799252","Type":"ContainerStarted","Data":"c55d0b3ef3dce1deaa6df76d43952c95558c3d394766c201f837bad69b12d9fc"} Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.948054 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-85jfr" podStartSLOduration=164.948028481 podStartE2EDuration="2m44.948028481s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.905206906 +0000 UTC m=+209.550076641" watchObservedRunningTime="2026-02-27 19:37:27.948028481 +0000 UTC m=+209.592898236" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.950579 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bvwws" podStartSLOduration=164.950562097 podStartE2EDuration="2m44.950562097s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.93490063 +0000 UTC m=+209.579770365" watchObservedRunningTime="2026-02-27 19:37:27.950562097 +0000 UTC m=+209.595431832" Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.957599 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" event={"ID":"202949b2-ad8d-451d-8ca3-0f94de759612","Type":"ContainerStarted","Data":"a198d8ecdcf65c16b80178a5c5169756f13fe1b8388ec2bd112231082240a7c9"} Feb 27 19:37:27 crc kubenswrapper[4839]: I0227 19:37:27.988161 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.020061 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48588: no serving certificate available for the kubelet" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.020541 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.020812 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.520785457 +0000 UTC m=+210.165655192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.020909 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.021612 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.521604932 +0000 UTC m=+210.166474667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.045192 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.046406 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pc7k8" podStartSLOduration=165.0463941 podStartE2EDuration="2m45.0463941s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:27.977499629 +0000 UTC m=+209.622369364" watchObservedRunningTime="2026-02-27 19:37:28.0463941 +0000 UTC m=+209.691263835" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.052575 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" event={"ID":"d51493b0-bd22-4a0e-9f54-c8a980289a4d","Type":"ContainerStarted","Data":"47a1f08fd61ead1ef66018d1a7caefaee8cf70fe4204581722841c0ffa7e6eb3"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.054488 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qlhjn" podStartSLOduration=165.05447047 podStartE2EDuration="2m45.05447047s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.009922064 +0000 UTC m=+209.654791799" watchObservedRunningTime="2026-02-27 19:37:28.05447047 +0000 UTC m=+209.699340205" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.054530 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.058976 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537016-rt648"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.081991 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kncvh" podStartSLOduration=165.081972599 podStartE2EDuration="2m45.081972599s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.077961389 +0000 UTC m=+209.722831124" watchObservedRunningTime="2026-02-27 19:37:28.081972599 +0000 UTC m=+209.726842334" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.087127 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" event={"ID":"f8b87c1a-1579-4415-8134-021c1d7c390e","Type":"ContainerStarted","Data":"c71fd756f0770a5b9c6ca7c59fb54e0e65d48d39b1bc1ef7e43fdb9f6bffbdf8"} Feb 27 19:37:28 crc kubenswrapper[4839]: W0227 19:37:28.092363 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ddc4982_fe8a_4121_9dc3_b71600c77843.slice/crio-dee61f3059d24d09033dc0a50fa7cfce325cbb664d391c11ae5cdb14d4a1fab1 WatchSource:0}: Error finding container dee61f3059d24d09033dc0a50fa7cfce325cbb664d391c11ae5cdb14d4a1fab1: Status 404 returned error can't find the container with id dee61f3059d24d09033dc0a50fa7cfce325cbb664d391c11ae5cdb14d4a1fab1 Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.105362 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-txbn6"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.111417 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vk2tn" podStartSLOduration=165.111402685 podStartE2EDuration="2m45.111402685s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.109279222 +0000 UTC m=+209.754148967" watchObservedRunningTime="2026-02-27 19:37:28.111402685 +0000 UTC m=+209.756272420" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.122709 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.123766 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.623746653 +0000 UTC m=+210.268616388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.180088 4839 generic.go:334] "Generic (PLEG): container finished" podID="e14522ed-1f77-462c-a5ec-caa1efe74a35" containerID="cd38ef31cc0e454c8b9037125f9155c40adc9b0d26d2e96d9496b9697f3a629c" exitCode=0 Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.180143 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" event={"ID":"e14522ed-1f77-462c-a5ec-caa1efe74a35","Type":"ContainerDied","Data":"cd38ef31cc0e454c8b9037125f9155c40adc9b0d26d2e96d9496b9697f3a629c"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.226347 4839 generic.go:334] "Generic (PLEG): container finished" podID="e7326984-1e3a-40e6-920d-ffdd7a2aaa89" containerID="2c368417a29811a5b8eaea49ecf54cac7cc7ef1df822178c8d51d508f58465fc" exitCode=0 Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.226447 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" event={"ID":"e7326984-1e3a-40e6-920d-ffdd7a2aaa89","Type":"ContainerDied","Data":"2c368417a29811a5b8eaea49ecf54cac7cc7ef1df822178c8d51d508f58465fc"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.227028 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.227775 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.231483 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.731468799 +0000 UTC m=+210.376338524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.262056 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48598: no serving certificate available for the kubelet" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.304883 4839 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.310490 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" event={"ID":"c1203271-c1c1-42e0-ab24-7403cc85726d","Type":"ContainerStarted","Data":"61a1aa7804a7e80eb19fe8384325404a6c90135fc7fcd11a359b1566c6becc16"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.320188 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6rfkk" event={"ID":"22eb69e2-54a0-4214-825e-0d3c5d3da62a","Type":"ContainerStarted","Data":"45b5ce5291ad3218cc5747bdbf8476435b51b64ba2f5d4f59093c8347e2492e6"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.328500 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.329688 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.829652812 +0000 UTC m=+210.474522537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.354628 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" event={"ID":"444f893d-0c15-4a0c-8805-38f51f8037f6","Type":"ContainerStarted","Data":"00225b5ca70cede98a30d6f423244d213c70a9520670778da79c44ca3376c42d"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.368634 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" event={"ID":"fb42fca6-045c-4c5d-8f0b-c29f40c746d0","Type":"ContainerStarted","Data":"0bbbbfd471f0f928d0b77a99d90dfd078ed893df9b9f3ebbab5588d67f0dca01"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.369662 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.371736 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" event={"ID":"b54c7b4f-365c-4088-bb22-94f6dc65c1c0","Type":"ContainerStarted","Data":"b960c689c0bdb78382072bc41cd24c52802e7e9632c48eae52958011b16650a3"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.373097 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" event={"ID":"df78f97a-5f5a-400e-a898-8c656da2c72f","Type":"ContainerStarted","Data":"f25cfc11a50c22da6e65e605eeee92dbd70245211dc1dcd9bcbe8871b0b3dbe5"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.387034 4839 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-bs9xm container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.387081 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" podUID="fb42fca6-045c-4c5d-8f0b-c29f40c746d0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.403363 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mflth" event={"ID":"93a37e84-0fbe-4825-a605-968b25a9eca2","Type":"ContainerStarted","Data":"23bb2b87add9f3c1fc025d1ab29cdd6f69c259ba322bb747bb8b8b475bae4243"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.436512 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.437150 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:28.937134282 +0000 UTC m=+210.582004097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.438282 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" podStartSLOduration=165.438260606 podStartE2EDuration="2m45.438260606s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.435520724 +0000 UTC m=+210.080390459" watchObservedRunningTime="2026-02-27 19:37:28.438260606 +0000 UTC m=+210.083130351" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.486330 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48602: no serving certificate available for the kubelet" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.491018 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-82v8f" event={"ID":"a4baae9f-93ba-40bf-9c28-01f13cdcaf43","Type":"ContainerStarted","Data":"56edbed701f168ea2522c00580c8a0c4edb07f5cfb9cfb7795d13b1dd0cc11c6"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.492195 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.503404 4839 patch_prober.go:28] interesting pod/downloads-7954f5f757-82v8f container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.503467 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-82v8f" podUID="a4baae9f-93ba-40bf-9c28-01f13cdcaf43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.536632 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-mflth" podStartSLOduration=165.536618964 podStartE2EDuration="2m45.536618964s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.533479351 +0000 UTC m=+210.178349086" watchObservedRunningTime="2026-02-27 19:37:28.536618964 +0000 UTC m=+210.181488699" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.537047 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5"] Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.537651 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.539092 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.039075247 +0000 UTC m=+210.683944982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.576091 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" event={"ID":"45305c05-bdb6-4b38-acdb-2fb7f472ed68","Type":"ContainerStarted","Data":"f9e429c9811484caab4527f461bf33df2ca32480d895dfac4827c8efe77804ec"} Feb 27 19:37:28 crc kubenswrapper[4839]: W0227 19:37:28.577304 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b927691_5423_4670_8443_c0a6a9d3bc13.slice/crio-8b8af4a16b1114faf54cee5113c79f1e5ee6c2877c6f71413fade4419f286dac WatchSource:0}: Error finding container 8b8af4a16b1114faf54cee5113c79f1e5ee6c2877c6f71413fade4419f286dac: Status 404 returned error can't find the container with id 8b8af4a16b1114faf54cee5113c79f1e5ee6c2877c6f71413fade4419f286dac Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.589011 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" event={"ID":"c34af3d9-4299-46e1-ba6d-56210fd9e5e2","Type":"ContainerStarted","Data":"cc8c01fee17a1b3232be45e7e72e16c7f588ea9a12873751aa7f7ad6183adea9"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.604550 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" event={"ID":"d5986cf5-a400-423c-916d-88260072d408","Type":"ContainerStarted","Data":"943dfc7ff05e9e85370759667ed960e2c2d7932eefc3faf579e2f2be7894ecba"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.609225 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" event={"ID":"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f","Type":"ContainerStarted","Data":"ba4eae8ebd5dd58239647b77a3d3360e02d443942f3d776beb67e245e4e26bf7"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.622908 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27r5h" podStartSLOduration=165.622893893 podStartE2EDuration="2m45.622893893s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.622323656 +0000 UTC m=+210.267193391" watchObservedRunningTime="2026-02-27 19:37:28.622893893 +0000 UTC m=+210.267763628" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.624651 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-82v8f" podStartSLOduration=165.624644185 podStartE2EDuration="2m45.624644185s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.592937531 +0000 UTC m=+210.237807266" watchObservedRunningTime="2026-02-27 19:37:28.624644185 +0000 UTC m=+210.269513920" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.638705 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.645654 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" event={"ID":"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c","Type":"ContainerStarted","Data":"d4f00d348616a22c368cfc002710a199d73219051ee3f2cb32e4f337e77c4005"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.646655 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.647066 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.147052752 +0000 UTC m=+210.791922527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.673998 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" podStartSLOduration=165.673976953 podStartE2EDuration="2m45.673976953s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.672069827 +0000 UTC m=+210.316939572" watchObservedRunningTime="2026-02-27 19:37:28.673976953 +0000 UTC m=+210.318846688" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.692212 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" event={"ID":"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2","Type":"ContainerStarted","Data":"d2395d253a18500bee67e573a7a2fb68bcaee9c7a072b88316100bf5fa0d0907"} Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.693005 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.699866 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:28 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:28 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:28 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.699909 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.700378 4839 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-clhjs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.700400 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.705326 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-9kxx2" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.707234 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qmwh7" podStartSLOduration=165.707218833 podStartE2EDuration="2m45.707218833s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.70578827 +0000 UTC m=+210.350657995" watchObservedRunningTime="2026-02-27 19:37:28.707218833 +0000 UTC m=+210.352088568" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.711307 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.743094 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.743288 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.243266276 +0000 UTC m=+210.888136011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.743896 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.746910 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.246896484 +0000 UTC m=+210.891766219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.765440 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" podStartSLOduration=165.765424546 podStartE2EDuration="2m45.765424546s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.742967717 +0000 UTC m=+210.387837462" watchObservedRunningTime="2026-02-27 19:37:28.765424546 +0000 UTC m=+210.410294281" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.815534 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" podStartSLOduration=165.815517857 podStartE2EDuration="2m45.815517857s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:28.771112605 +0000 UTC m=+210.415982340" watchObservedRunningTime="2026-02-27 19:37:28.815517857 +0000 UTC m=+210.460387592" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.845017 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.848901 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.34888031 +0000 UTC m=+210.993750045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.852328 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.852772 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.352734885 +0000 UTC m=+210.997604640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.895316 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48604: no serving certificate available for the kubelet" Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.953618 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.953812 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.453776743 +0000 UTC m=+211.098646478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:28 crc kubenswrapper[4839]: I0227 19:37:28.954175 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:28 crc kubenswrapper[4839]: E0227 19:37:28.954445 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.454434133 +0000 UTC m=+211.099303868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.095420 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.096041 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.596022888 +0000 UTC m=+211.240892623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.137499 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.196985 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.197514 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.697500299 +0000 UTC m=+211.342370034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.303714 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.304016 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.80400069 +0000 UTC m=+211.448870425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.414497 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.415023 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:29.915010414 +0000 UTC m=+211.559880149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.515204 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.515473 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.015423574 +0000 UTC m=+211.660293319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.519281 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.519606 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.019590218 +0000 UTC m=+211.664459953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.624201 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.624763 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.124722958 +0000 UTC m=+211.769592703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.644112 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48612: no serving certificate available for the kubelet" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.680272 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:29 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:29 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:29 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.680314 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.725086 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.725481 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.225463327 +0000 UTC m=+211.870333062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.730911 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.731019 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:37:29 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:37:29 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmg87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537016-rt648_openshift-infra(95e89fb2-3271-46bd-878b-6ff642b17214): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:37:29 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.733946 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.737163 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" event={"ID":"6ddc4982-fe8a-4121-9dc3-b71600c77843","Type":"ContainerStarted","Data":"2b006fdabbd45e81efbf11ab9e0fa90080f095005da74e90c1d56a8f078c0c48"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.737210 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" event={"ID":"6ddc4982-fe8a-4121-9dc3-b71600c77843","Type":"ContainerStarted","Data":"dee61f3059d24d09033dc0a50fa7cfce325cbb664d391c11ae5cdb14d4a1fab1"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.766374 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zsp52" event={"ID":"00d587ff-502e-42ed-9d00-02f639086a6d","Type":"ContainerStarted","Data":"6b5eb150e71cd0168d675ab8a5bd5ae91c74c3c4e34fc7b4c0f1c06f3fc7063b"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.796052 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" event={"ID":"c1203271-c1c1-42e0-ab24-7403cc85726d","Type":"ContainerStarted","Data":"8fb24ea9a420f6fd4a11ede93a8e40d3f203bb7de0ecf0fd36b104c1a8fff3ba"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.796111 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" event={"ID":"c1203271-c1c1-42e0-ab24-7403cc85726d","Type":"ContainerStarted","Data":"55c7e5aa5f9a80f1d5f5a0b34d52ba0692219894cc3544754f8f1d29326e3793"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.816911 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zsp52" podStartSLOduration=7.816891779 podStartE2EDuration="7.816891779s" podCreationTimestamp="2026-02-27 19:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:29.815050784 +0000 UTC m=+211.459920519" watchObservedRunningTime="2026-02-27 19:37:29.816891779 +0000 UTC m=+211.461761504" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.829365 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.830151 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.330135273 +0000 UTC m=+211.975004998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.851642 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" event={"ID":"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f","Type":"ContainerStarted","Data":"84309637159d22bf750ad6af27110b9c0a0650c15887e32c716fa155d56ce725"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.855023 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gxn9q" podStartSLOduration=166.855006134 podStartE2EDuration="2m46.855006134s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:29.852150079 +0000 UTC m=+211.497019814" watchObservedRunningTime="2026-02-27 19:37:29.855006134 +0000 UTC m=+211.499875869" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.882804 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" event={"ID":"710bae1b-9785-4c51-a128-f91f4e962113","Type":"ContainerStarted","Data":"0cca729cbf48ebf149ecc7ba4ad73d6c3e6d142fd1d2fc143aacd9006d73ab60"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.882860 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.908291 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6rfkk" event={"ID":"22eb69e2-54a0-4214-825e-0d3c5d3da62a","Type":"ContainerStarted","Data":"1c0173ad1cd9ac0e47ef688b4477ee8a48b14964560ae51e70b927612ffba4e8"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.916480 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" event={"ID":"22e4cd58-dab8-41fa-9cae-323ce54a4b0d","Type":"ContainerStarted","Data":"f8f5b9a0cdd9418eef03587c634d783e2d27654dd6755f1c3396c8ae02a86f7f"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.916522 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" event={"ID":"22e4cd58-dab8-41fa-9cae-323ce54a4b0d","Type":"ContainerStarted","Data":"e6a422d806d08f6036a6a80a46e08470014f2bd510b23c2ff782dc338e1aba60"} Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.934738 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:29 crc kubenswrapper[4839]: E0227 19:37:29.936855 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.43684194 +0000 UTC m=+212.081711675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.969475 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" podStartSLOduration=166.969458931 podStartE2EDuration="2m46.969458931s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:29.957061552 +0000 UTC m=+211.601931287" watchObservedRunningTime="2026-02-27 19:37:29.969458931 +0000 UTC m=+211.614328666" Feb 27 19:37:29 crc kubenswrapper[4839]: I0227 19:37:29.970827 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbbmg"] Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.031421 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" event={"ID":"444f893d-0c15-4a0c-8805-38f51f8037f6","Type":"ContainerStarted","Data":"5e27743bdf24244a9939de4ec1a0a76be06250907a0f3897138f14b10867dc31"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.039354 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.039827 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.539806945 +0000 UTC m=+212.184676680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.077384 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" event={"ID":"e14522ed-1f77-462c-a5ec-caa1efe74a35","Type":"ContainerStarted","Data":"cab4d518c13c2ebcdec37cdb953ae6ffc98bc7c24a8fa2992201ca6d281e9fb0"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.078114 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-pgbmd" podStartSLOduration=167.078093355 podStartE2EDuration="2m47.078093355s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.077199318 +0000 UTC m=+211.722069053" watchObservedRunningTime="2026-02-27 19:37:30.078093355 +0000 UTC m=+211.722963090" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.090133 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" event={"ID":"1c376125-89b2-43d4-970a-c13ff4fbed59","Type":"ContainerStarted","Data":"c00344be34b9c76bd49b03e635e9a6ba789eac51e1c43289df8c20d306662560"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.138321 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d"] Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.140401 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6rfkk" podStartSLOduration=8.140377559 podStartE2EDuration="8.140377559s" podCreationTimestamp="2026-02-27 19:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.139906795 +0000 UTC m=+211.784776530" watchObservedRunningTime="2026-02-27 19:37:30.140377559 +0000 UTC m=+211.785247294" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.141721 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.166315 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.666299101 +0000 UTC m=+212.311168836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.173146 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" event={"ID":"a783a0e0-5fb8-4542-b61c-bdd54fc3a331","Type":"ContainerStarted","Data":"4d16a5f9091fe8f6be0ac0f7877eaca752b2b9f72030f9cebc1711ebe02240b1"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.173204 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" event={"ID":"a783a0e0-5fb8-4542-b61c-bdd54fc3a331","Type":"ContainerStarted","Data":"67046890ae6f4dffef42195175fdc93694cd5f9d6ceb39e552999f1dc8597317"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.180128 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" event={"ID":"b54c7b4f-365c-4088-bb22-94f6dc65c1c0","Type":"ContainerStarted","Data":"3548f983ca206da751844a78427c7138eeece06e66f7560628752d7b988d3961"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.181715 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" event={"ID":"df78f97a-5f5a-400e-a898-8c656da2c72f","Type":"ContainerStarted","Data":"0d25e108b3f81aeb4da541738c3acba99dd4f3a97f93ace932a5aec7f5550695"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.182061 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.198946 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" event={"ID":"ce8c6777-cb48-40db-b97c-e79398a0059e","Type":"ContainerStarted","Data":"6faf17442103e333218953355a51ebf8849a1cd36ff448d01ab30b5b826dfdec"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.206033 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.208642 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-t69c7" event={"ID":"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89","Type":"ContainerStarted","Data":"86c77d123cfe521abc53cb244e092103637c8d8b2f2c77522ece1b85e01d11f8"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.208699 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-t69c7" event={"ID":"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89","Type":"ContainerStarted","Data":"ba828f681fd26ecb75c6a37222efc9da81c37d4b76c35cd4c915aeda0b27f442"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.210551 4839 patch_prober.go:28] interesting pod/apiserver-76f77b778f-qfgcq container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.210603 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" podUID="e14522ed-1f77-462c-a5ec-caa1efe74a35" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.211108 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.216410 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" event={"ID":"fb42fca6-045c-4c5d-8f0b-c29f40c746d0","Type":"ContainerStarted","Data":"824473da1cd36806f57dac82656103d0f66f9bbfb89c5a4839a39cf8f1ced603"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253159 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253554 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253600 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-67h5d" podStartSLOduration=167.25358357 podStartE2EDuration="2m47.25358357s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.179234646 +0000 UTC m=+211.824104371" watchObservedRunningTime="2026-02-27 19:37:30.25358357 +0000 UTC m=+211.898453305" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253646 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253704 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253743 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.253985 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-729pl" podStartSLOduration=167.253981871 podStartE2EDuration="2m47.253981871s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.252750365 +0000 UTC m=+211.897620110" watchObservedRunningTime="2026-02-27 19:37:30.253981871 +0000 UTC m=+211.898851606" Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.254365 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.754350612 +0000 UTC m=+212.399220347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.268379 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.276647 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.277245 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.283618 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.293058 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" event={"ID":"d8a6578e-ff7c-4cb6-a051-2dabed753669","Type":"ContainerStarted","Data":"5b99f3beb91a6b937bf6089f11db69d5b8f9b4626a7be1789b7126c84c35639c"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.325958 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8qr2g" event={"ID":"c34af3d9-4299-46e1-ba6d-56210fd9e5e2","Type":"ContainerStarted","Data":"1bb3ffe870967e3c926901ad78e093d2d27ba457b51f479a61711ee238b608d0"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.331451 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" podStartSLOduration=167.331422767 podStartE2EDuration="2m47.331422767s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.301135005 +0000 UTC m=+211.946004740" watchObservedRunningTime="2026-02-27 19:37:30.331422767 +0000 UTC m=+211.976292522" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.337195 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" event={"ID":"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a","Type":"ContainerStarted","Data":"f445b368ff8cac1c7abb95066be779938191fb124314c3e9015103216d55c8a5"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.354580 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.354618 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.354920 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.854908856 +0000 UTC m=+212.499778591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.355986 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537016-rt648" event={"ID":"95e89fb2-3271-46bd-878b-6ff642b17214","Type":"ContainerStarted","Data":"58c003c295f585d416eb192851977ffecd334166767f3ebf3f304da6448aa3f7"} Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.362278 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.363171 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b-metrics-certs\") pod \"network-metrics-daemon-m6927\" (UID: \"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b\") " pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.376591 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" event={"ID":"4a9a60f2-147d-49c1-beaa-7f13ad42737a","Type":"ContainerStarted","Data":"8cbacb3d49edfbd3970bc4be4e4bb034ecca8f02451e6a44c4311020fdfcea09"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.376634 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" event={"ID":"4a9a60f2-147d-49c1-beaa-7f13ad42737a","Type":"ContainerStarted","Data":"70047ecc0233102e056b9af78856e05aeb58825524736909b066d22e5e19c043"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.377545 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.381863 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bs9xm" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.382671 4839 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2cx5s container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.382754 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" podUID="4a9a60f2-147d-49c1-beaa-7f13ad42737a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.391552 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" event={"ID":"d18ed308-bc67-4689-8926-3ceb3b799252","Type":"ContainerStarted","Data":"fa5854291b8d4e48c7b119adf2e0b65d24d0f32f92156eb0332dc331b83206ff"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.392123 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.393254 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" podStartSLOduration=167.393235616 podStartE2EDuration="2m47.393235616s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.391849385 +0000 UTC m=+212.036719140" watchObservedRunningTime="2026-02-27 19:37:30.393235616 +0000 UTC m=+212.038105351" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.415294 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.420056 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" event={"ID":"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2","Type":"ContainerStarted","Data":"a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.421069 4839 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-clhjs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.421110 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.435766 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6927" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.443373 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.449280 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" event={"ID":"365ad46d-8d32-45b7-bcbe-e4b771babb9a","Type":"ContainerStarted","Data":"8175c7725759e76d37e7bd8dcfa2fd1a196c3c6fa805251d2e580321723a678b"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.454537 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-vgnmn" podStartSLOduration=167.454517731 podStartE2EDuration="2m47.454517731s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.448854702 +0000 UTC m=+212.093724467" watchObservedRunningTime="2026-02-27 19:37:30.454517731 +0000 UTC m=+212.099387466" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.455546 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.456855 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:30.956807239 +0000 UTC m=+212.601676974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.510002 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mflth" event={"ID":"93a37e84-0fbe-4825-a605-968b25a9eca2","Type":"ContainerStarted","Data":"6b314d9846234367a8e1c2e4ecc3dff14d42db5b04849172cbdce144282e27dc"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.522973 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v2gv2" podStartSLOduration=167.522956948 podStartE2EDuration="2m47.522956948s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.490539803 +0000 UTC m=+212.135409548" watchObservedRunningTime="2026-02-27 19:37:30.522956948 +0000 UTC m=+212.167826683" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.557039 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" event={"ID":"4b927691-5423-4670-8443-c0a6a9d3bc13","Type":"ContainerStarted","Data":"8b8af4a16b1114faf54cee5113c79f1e5ee6c2877c6f71413fade4419f286dac"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.563326 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.563852 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.063817154 +0000 UTC m=+212.708686979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.571748 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kfdf5" podStartSLOduration=167.57173227 podStartE2EDuration="2m47.57173227s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.527110912 +0000 UTC m=+212.171980647" watchObservedRunningTime="2026-02-27 19:37:30.57173227 +0000 UTC m=+212.216602005" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.581103 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" event={"ID":"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4","Type":"ContainerStarted","Data":"173265a561cb9e14e933227ec29917fe7f2672aad54285b19ae632d129be478c"} Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.586779 4839 patch_prober.go:28] interesting pod/downloads-7954f5f757-82v8f container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.586815 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-82v8f" podUID="a4baae9f-93ba-40bf-9c28-01f13cdcaf43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.617563 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" podStartSLOduration=167.617543514 podStartE2EDuration="2m47.617543514s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.572848103 +0000 UTC m=+212.217717838" watchObservedRunningTime="2026-02-27 19:37:30.617543514 +0000 UTC m=+212.262413259" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.664388 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.664959 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.164944335 +0000 UTC m=+212.809814070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.691301 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:30 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:30 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:30 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.691355 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.737458 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" podStartSLOduration=167.737439883 podStartE2EDuration="2m47.737439883s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.715185091 +0000 UTC m=+212.360054826" watchObservedRunningTime="2026-02-27 19:37:30.737439883 +0000 UTC m=+212.382309618" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.766084 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.769142 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" podStartSLOduration=167.769128357 podStartE2EDuration="2m47.769128357s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.766926801 +0000 UTC m=+212.411796536" watchObservedRunningTime="2026-02-27 19:37:30.769128357 +0000 UTC m=+212.413998092" Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.774201 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" podStartSLOduration=167.774184277 podStartE2EDuration="2m47.774184277s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.739149704 +0000 UTC m=+212.384019449" watchObservedRunningTime="2026-02-27 19:37:30.774184277 +0000 UTC m=+212.419054032" Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.778092 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.278078253 +0000 UTC m=+212.922947978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.872351 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.872618 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.372603137 +0000 UTC m=+213.017472872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.978448 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:30 crc kubenswrapper[4839]: E0227 19:37:30.979085 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.479068717 +0000 UTC m=+213.123938452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:30 crc kubenswrapper[4839]: I0227 19:37:30.991023 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48618: no serving certificate available for the kubelet" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.081171 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.081636 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.58162021 +0000 UTC m=+213.226489945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.183462 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.183793 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" podStartSLOduration=168.183774521 podStartE2EDuration="2m48.183774521s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:30.804073437 +0000 UTC m=+212.448943172" watchObservedRunningTime="2026-02-27 19:37:31.183774521 +0000 UTC m=+212.828644256" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.183838 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.683823943 +0000 UTC m=+213.328693678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: W0227 19:37:31.208898 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-a37f7fe04108da9a5384998e96247cf047e8492b8307f6bd3bde230c11726a95 WatchSource:0}: Error finding container a37f7fe04108da9a5384998e96247cf047e8492b8307f6bd3bde230c11726a95: Status 404 returned error can't find the container with id a37f7fe04108da9a5384998e96247cf047e8492b8307f6bd3bde230c11726a95 Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.247068 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gbl86" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.284981 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.285277 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.785257052 +0000 UTC m=+213.430126787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.373148 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m6927"] Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.388879 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.389330 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.88931016 +0000 UTC m=+213.534179885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.489732 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.489855 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.989835113 +0000 UTC m=+213.634704848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.490072 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.490396 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:31.990385919 +0000 UTC m=+213.635255654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.590647 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.590829 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.090806489 +0000 UTC m=+213.735676234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.591321 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.591601 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.091589732 +0000 UTC m=+213.736459467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.598565 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0f066538c2a18164966b351e2f1626f8bf5ad18a20d113d78e4ea8ee255edf3c"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.598613 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a37f7fe04108da9a5384998e96247cf047e8492b8307f6bd3bde230c11726a95"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.604495 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s7t9g" event={"ID":"d8a6578e-ff7c-4cb6-a051-2dabed753669","Type":"ContainerStarted","Data":"163833c4f37d750b146bcedc94e400233722f7e370cd8117385d1ff9ee5e193e"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.609182 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m6927" event={"ID":"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b","Type":"ContainerStarted","Data":"725eed73b4787ef6e6fdc5382d03aecc2c9a7aed396b709884ff034b2def77f3"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.612866 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4x5z" event={"ID":"0ec24bbd-8a94-4b3f-ac19-40eda8ab4d7a","Type":"ContainerStarted","Data":"46c41b492ec49ad281f25f240cf0306fa976e30f4994d5815af3fbcee7b20ba5"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.615820 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1537af71922884451fdee29779d86675668b8f3c37493ff678f54d48706ec178"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.615847 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0676e17f559027306f6183137c3c70027eae0e92fee958b94bc9554f17b889f1"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.616170 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.617891 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" event={"ID":"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4","Type":"ContainerStarted","Data":"e5b6d1947cbbba92e82812e5e03b33d3ff8be397cc3c683daec0e8e416ed02e4"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.620006 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" event={"ID":"e7326984-1e3a-40e6-920d-ffdd7a2aaa89","Type":"ContainerStarted","Data":"6cce39532a687298dd8d8fac498d82fbeb593bdd9dcdfd96c0f4b0436d315142"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.622385 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-w4ph9" event={"ID":"365ad46d-8d32-45b7-bcbe-e4b771babb9a","Type":"ContainerStarted","Data":"cb855b87a42c3df4980ce3ccb821a46c93ada699ada3f7c3dc6743a289d3df0b"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.625996 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" event={"ID":"4b927691-5423-4670-8443-c0a6a9d3bc13","Type":"ContainerStarted","Data":"12ad453b448f74de4d775ae89cccfc7a64b1bb9268c9e9e368e2e7b88186e9a1"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.627967 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" event={"ID":"c6d6ffca-5f53-4f1d-b942-8b242b2b7d4f","Type":"ContainerStarted","Data":"453ecd0f8d32bc39647dfa661726502fbb5e62da4761fd459e4713e88b6f5a5f"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.634579 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" event={"ID":"e14522ed-1f77-462c-a5ec-caa1efe74a35","Type":"ContainerStarted","Data":"45c03e283325756651b4d8958a3b40befc67c4377436ac543b61911a1d0e0f82"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.638399 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" event={"ID":"d18ed308-bc67-4689-8926-3ceb3b799252","Type":"ContainerStarted","Data":"19fdaec9c93967d63b137ba35a4ce77417a2af3df487996b5e8811c192e08652"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.640757 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-t69c7" event={"ID":"aaca7144-ec64-4c3a-8fcd-b77b9d51fb89","Type":"ContainerStarted","Data":"fc79bfc55ec9f7cf192b9920a40073cf98d0e3e6b397ede9b649bae7d6b5d701"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.641094 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.655983 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2db37a2979b685d99c3c3aaa851181aee5e66eaad8546faa33ffd706669b1d83"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.662564 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" event={"ID":"df78f97a-5f5a-400e-a898-8c656da2c72f","Type":"ContainerStarted","Data":"6682e9b6a3c4e0b098c06a87f189218785576249488f75b96fea2bdef391f694"} Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.665469 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" podUID="9ef6e180-79df-40ee-bf2b-96332e35c052" containerName="controller-manager" containerID="cri-o://0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949" gracePeriod=30 Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.667070 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" podUID="b6f71939-65e0-4459-9c58-211f5cafd259" containerName="route-controller-manager" containerID="cri-o://10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424" gracePeriod=30 Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.669779 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.669815 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" podStartSLOduration=168.66979976 podStartE2EDuration="2m48.66979976s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:31.646094115 +0000 UTC m=+213.290963850" watchObservedRunningTime="2026-02-27 19:37:31.66979976 +0000 UTC m=+213.314669495" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.676732 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-stz5c"] Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.677872 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.681470 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.686281 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-stz5c"] Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.686936 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:31 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:31 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:31 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.686991 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.687267 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.696887 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.697053 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.197024191 +0000 UTC m=+213.841893936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.697447 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.699416 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.199403562 +0000 UTC m=+213.844273297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.739843 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" podStartSLOduration=168.739821325 podStartE2EDuration="2m48.739821325s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:31.735215878 +0000 UTC m=+213.380085603" watchObservedRunningTime="2026-02-27 19:37:31.739821325 +0000 UTC m=+213.384691080" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.767716 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtjpz" podStartSLOduration=168.767692855 podStartE2EDuration="2m48.767692855s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:31.761236523 +0000 UTC m=+213.406106258" watchObservedRunningTime="2026-02-27 19:37:31.767692855 +0000 UTC m=+213.412562590" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.799050 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.799460 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-catalog-content\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.799875 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8btrd\" (UniqueName: \"kubernetes.io/projected/ff0a141c-3800-4c84-b1d6-17ec909d5d27-kube-api-access-8btrd\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.799947 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-utilities\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.801415 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.301387218 +0000 UTC m=+213.946256983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.881870 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-fszhl" podStartSLOduration=168.881847073 podStartE2EDuration="2m48.881847073s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:31.871177366 +0000 UTC m=+213.516047111" watchObservedRunningTime="2026-02-27 19:37:31.881847073 +0000 UTC m=+213.526716818" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.901053 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8btrd\" (UniqueName: \"kubernetes.io/projected/ff0a141c-3800-4c84-b1d6-17ec909d5d27-kube-api-access-8btrd\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.901098 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-utilities\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.901173 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-catalog-content\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.901202 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:31 crc kubenswrapper[4839]: E0227 19:37:31.901518 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.401502839 +0000 UTC m=+214.046372574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.902339 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-utilities\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.902615 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-catalog-content\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.904796 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d2p75"] Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.906237 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.907813 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.919894 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2p75"] Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.925716 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-t69c7" podStartSLOduration=9.922315138 podStartE2EDuration="9.922315138s" podCreationTimestamp="2026-02-27 19:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:31.907122146 +0000 UTC m=+213.551991891" watchObservedRunningTime="2026-02-27 19:37:31.922315138 +0000 UTC m=+213.567184873" Feb 27 19:37:31 crc kubenswrapper[4839]: I0227 19:37:31.931637 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8btrd\" (UniqueName: \"kubernetes.io/projected/ff0a141c-3800-4c84-b1d6-17ec909d5d27-kube-api-access-8btrd\") pod \"certified-operators-stz5c\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.002185 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.002456 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5wrd\" (UniqueName: \"kubernetes.io/projected/b471c70a-15fe-4b5f-aa6f-354848925ace-kube-api-access-z5wrd\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.002531 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-utilities\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.002572 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-catalog-content\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.002725 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.502705421 +0000 UTC m=+214.147575166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.005490 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.077864 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5p9bn"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.078939 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.103357 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-utilities\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.103410 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.103441 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-catalog-content\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.103496 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5wrd\" (UniqueName: \"kubernetes.io/projected/b471c70a-15fe-4b5f-aa6f-354848925ace-kube-api-access-z5wrd\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.104192 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-utilities\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.104483 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.604470741 +0000 UTC m=+214.249340476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.104901 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-catalog-content\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.113364 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5p9bn"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.154807 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5wrd\" (UniqueName: \"kubernetes.io/projected/b471c70a-15fe-4b5f-aa6f-354848925ace-kube-api-access-z5wrd\") pod \"community-operators-d2p75\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.209382 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.209919 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-catalog-content\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.209958 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-utilities\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.209981 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvsn8\" (UniqueName: \"kubernetes.io/projected/8cc72c87-b6d6-4384-ac6e-fb12af841f66-kube-api-access-mvsn8\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.210147 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.710127326 +0000 UTC m=+214.354997061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.260493 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.271944 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgzcw"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.273300 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.305306 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgzcw"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.314795 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-utilities\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.314833 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvsn8\" (UniqueName: \"kubernetes.io/projected/8cc72c87-b6d6-4384-ac6e-fb12af841f66-kube-api-access-mvsn8\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.314882 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.314961 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-catalog-content\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.315415 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-catalog-content\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.315989 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-utilities\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.319786 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.81976317 +0000 UTC m=+214.464632965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.348298 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvsn8\" (UniqueName: \"kubernetes.io/projected/8cc72c87-b6d6-4384-ac6e-fb12af841f66-kube-api-access-mvsn8\") pod \"certified-operators-5p9bn\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.394857 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.411151 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.421723 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.421950 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-utilities\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.421991 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm9fn\" (UniqueName: \"kubernetes.io/projected/0a4d0578-88dd-4754-81fd-ae28a709b698-kube-api-access-zm9fn\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.422021 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-catalog-content\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.422150 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:32.922131448 +0000 UTC m=+214.567001183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.436608 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs"] Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.436860 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef6e180-79df-40ee-bf2b-96332e35c052" containerName="controller-manager" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.436875 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef6e180-79df-40ee-bf2b-96332e35c052" containerName="controller-manager" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.436999 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef6e180-79df-40ee-bf2b-96332e35c052" containerName="controller-manager" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.437408 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.458910 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.486145 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.523123 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ef6e180-79df-40ee-bf2b-96332e35c052-serving-cert\") pod \"9ef6e180-79df-40ee-bf2b-96332e35c052\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.524584 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-config\") pod \"9ef6e180-79df-40ee-bf2b-96332e35c052\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.524631 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-client-ca\") pod \"9ef6e180-79df-40ee-bf2b-96332e35c052\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.524831 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wts7z\" (UniqueName: \"kubernetes.io/projected/9ef6e180-79df-40ee-bf2b-96332e35c052-kube-api-access-wts7z\") pod \"9ef6e180-79df-40ee-bf2b-96332e35c052\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.524865 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-proxy-ca-bundles\") pod \"9ef6e180-79df-40ee-bf2b-96332e35c052\" (UID: \"9ef6e180-79df-40ee-bf2b-96332e35c052\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525145 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-proxy-ca-bundles\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525195 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xx8n\" (UniqueName: \"kubernetes.io/projected/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-kube-api-access-2xx8n\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525231 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-utilities\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525256 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-client-ca\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525293 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm9fn\" (UniqueName: \"kubernetes.io/projected/0a4d0578-88dd-4754-81fd-ae28a709b698-kube-api-access-zm9fn\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525319 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-catalog-content\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525366 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525397 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-config\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525418 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-serving-cert\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525897 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-client-ca" (OuterVolumeSpecName: "client-ca") pod "9ef6e180-79df-40ee-bf2b-96332e35c052" (UID: "9ef6e180-79df-40ee-bf2b-96332e35c052"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.525948 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-config" (OuterVolumeSpecName: "config") pod "9ef6e180-79df-40ee-bf2b-96332e35c052" (UID: "9ef6e180-79df-40ee-bf2b-96332e35c052"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.526483 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9ef6e180-79df-40ee-bf2b-96332e35c052" (UID: "9ef6e180-79df-40ee-bf2b-96332e35c052"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.526741 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-catalog-content\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.526866 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-utilities\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.527192 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.027175705 +0000 UTC m=+214.672045550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.531073 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ef6e180-79df-40ee-bf2b-96332e35c052-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9ef6e180-79df-40ee-bf2b-96332e35c052" (UID: "9ef6e180-79df-40ee-bf2b-96332e35c052"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.544460 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef6e180-79df-40ee-bf2b-96332e35c052-kube-api-access-wts7z" (OuterVolumeSpecName: "kube-api-access-wts7z") pod "9ef6e180-79df-40ee-bf2b-96332e35c052" (UID: "9ef6e180-79df-40ee-bf2b-96332e35c052"). InnerVolumeSpecName "kube-api-access-wts7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.554590 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cx5s" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.556782 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm9fn\" (UniqueName: \"kubernetes.io/projected/0a4d0578-88dd-4754-81fd-ae28a709b698-kube-api-access-zm9fn\") pod \"community-operators-cgzcw\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.593897 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.626481 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f71939-65e0-4459-9c58-211f5cafd259-serving-cert\") pod \"b6f71939-65e0-4459-9c58-211f5cafd259\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.626640 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.626702 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m97gl\" (UniqueName: \"kubernetes.io/projected/b6f71939-65e0-4459-9c58-211f5cafd259-kube-api-access-m97gl\") pod \"b6f71939-65e0-4459-9c58-211f5cafd259\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.626799 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-client-ca\") pod \"b6f71939-65e0-4459-9c58-211f5cafd259\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.626846 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-config\") pod \"b6f71939-65e0-4459-9c58-211f5cafd259\" (UID: \"b6f71939-65e0-4459-9c58-211f5cafd259\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.626997 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-proxy-ca-bundles\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627037 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xx8n\" (UniqueName: \"kubernetes.io/projected/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-kube-api-access-2xx8n\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627070 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-client-ca\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627144 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-config\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627166 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-serving-cert\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627207 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ef6e180-79df-40ee-bf2b-96332e35c052-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627221 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627233 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627243 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wts7z\" (UniqueName: \"kubernetes.io/projected/9ef6e180-79df-40ee-bf2b-96332e35c052-kube-api-access-wts7z\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.627254 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ef6e180-79df-40ee-bf2b-96332e35c052-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.628837 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.128813061 +0000 UTC m=+214.773682796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.630271 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-client-ca\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.630966 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-proxy-ca-bundles\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.631102 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-config\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.631767 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-client-ca" (OuterVolumeSpecName: "client-ca") pod "b6f71939-65e0-4459-9c58-211f5cafd259" (UID: "b6f71939-65e0-4459-9c58-211f5cafd259"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.631841 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-serving-cert\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.631995 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-config" (OuterVolumeSpecName: "config") pod "b6f71939-65e0-4459-9c58-211f5cafd259" (UID: "b6f71939-65e0-4459-9c58-211f5cafd259"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.640189 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f71939-65e0-4459-9c58-211f5cafd259-kube-api-access-m97gl" (OuterVolumeSpecName: "kube-api-access-m97gl") pod "b6f71939-65e0-4459-9c58-211f5cafd259" (UID: "b6f71939-65e0-4459-9c58-211f5cafd259"). InnerVolumeSpecName "kube-api-access-m97gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.640522 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6f71939-65e0-4459-9c58-211f5cafd259-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b6f71939-65e0-4459-9c58-211f5cafd259" (UID: "b6f71939-65e0-4459-9c58-211f5cafd259"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.672397 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xx8n\" (UniqueName: \"kubernetes.io/projected/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-kube-api-access-2xx8n\") pod \"controller-manager-9cdb8d79d-zhqrs\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.678171 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:32 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:32 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:32 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.678226 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.717162 4839 generic.go:334] "Generic (PLEG): container finished" podID="b6f71939-65e0-4459-9c58-211f5cafd259" containerID="10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424" exitCode=0 Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.717259 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" event={"ID":"b6f71939-65e0-4459-9c58-211f5cafd259","Type":"ContainerDied","Data":"10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424"} Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.717290 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" event={"ID":"b6f71939-65e0-4459-9c58-211f5cafd259","Type":"ContainerDied","Data":"5b92219f064e325beab6765f8e32961736356c11e77b621078b21bc7cb4e06c6"} Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.717308 4839 scope.go:117] "RemoveContainer" containerID="10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.717446 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.728357 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.728434 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f71939-65e0-4459-9c58-211f5cafd259-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.728449 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m97gl\" (UniqueName: \"kubernetes.io/projected/b6f71939-65e0-4459-9c58-211f5cafd259-kube-api-access-m97gl\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.728463 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.728474 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f71939-65e0-4459-9c58-211f5cafd259-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.728778 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.228763597 +0000 UTC m=+214.873633332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.750921 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"68857f81105258b28293b645daed28e6f62361d8a3302a4c30a89ab08bba253f"} Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.765604 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m6927" event={"ID":"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b","Type":"ContainerStarted","Data":"40ff00a1b44aec13d8824560e0775afdeb4f35e36b847e0611ff26d3dea4ed5b"} Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.771432 4839 scope.go:117] "RemoveContainer" containerID="10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.772282 4839 generic.go:334] "Generic (PLEG): container finished" podID="9ef6e180-79df-40ee-bf2b-96332e35c052" containerID="0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949" exitCode=0 Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.774291 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424\": container with ID starting with 10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424 not found: ID does not exist" containerID="10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.774317 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424"} err="failed to get container status \"10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424\": rpc error: code = NotFound desc = could not find container \"10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424\": container with ID starting with 10ce67110dbd5bf1b76e83ef34e39c1de40e8f1f83fbabd8536d52eb1e957424 not found: ID does not exist" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.774359 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.775734 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" event={"ID":"9ef6e180-79df-40ee-bf2b-96332e35c052","Type":"ContainerDied","Data":"0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949"} Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.775786 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbbmg" event={"ID":"9ef6e180-79df-40ee-bf2b-96332e35c052","Type":"ContainerDied","Data":"79446c82f49ca00e64f617e1bf04a1499f6134e9fa86aaa22d9e46576f2196b4"} Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.775808 4839 scope.go:117] "RemoveContainer" containerID="0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.786685 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.839322 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.839628 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.839956 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.339935936 +0000 UTC m=+214.984805671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.840057 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.840764 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.340752691 +0000 UTC m=+214.985622426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.839326 4839 scope.go:117] "RemoveContainer" containerID="0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.843639 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bpkff" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.848155 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-stz5c"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.853778 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2p75"] Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.858600 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949\": container with ID starting with 0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949 not found: ID does not exist" containerID="0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.858686 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949"} err="failed to get container status \"0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949\": rpc error: code = NotFound desc = could not find container \"0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949\": container with ID starting with 0216b1d3acdd00de5c6c08fad79f9126ee50d2174642a54072e55de3a1b5a949 not found: ID does not exist" Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.878949 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.905172 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gm67d"] Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.942870 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.943268 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.443218571 +0000 UTC m=+215.088088306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:32 crc kubenswrapper[4839]: I0227 19:37:32.948785 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:32 crc kubenswrapper[4839]: E0227 19:37:32.949242 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.44922456 +0000 UTC m=+215.094094385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.053616 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f71939-65e0-4459-9c58-211f5cafd259" path="/var/lib/kubelet/pods/b6f71939-65e0-4459-9c58-211f5cafd259/volumes" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.054346 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbbmg"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.054366 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbbmg"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.058306 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.058673 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.558644478 +0000 UTC m=+215.203514213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.136228 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5p9bn"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.160523 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.160896 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.660884881 +0000 UTC m=+215.305754616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.242035 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgzcw"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.262401 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.262697 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.762681252 +0000 UTC m=+215.407550987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.302020 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.365575 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.373734 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.866158032 +0000 UTC m=+215.511027777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.467330 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.467538 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.9675183 +0000 UTC m=+215.612388035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.467909 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.468268 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:33.968257602 +0000 UTC m=+215.613127337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.509722 4839 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.569897 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.570123 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.070094094 +0000 UTC m=+215.714963829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.570330 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.570770 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.070752203 +0000 UTC m=+215.715621938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.578505 4839 ???:1] "http: TLS handshake error from 192.168.126.11:48630: no serving certificate available for the kubelet" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.671337 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.671470 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.171441101 +0000 UTC m=+215.816310836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.671584 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.671884 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.171873574 +0000 UTC m=+215.816743309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.678805 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:33 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:33 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:33 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.678863 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.773259 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.773638 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.273614553 +0000 UTC m=+215.918484288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.812529 4839 generic.go:334] "Generic (PLEG): container finished" podID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerID="af4d438849daa5d4e132ae67e3659b425f69071f5b4262fbdd5a5f7837f9cc04" exitCode=0 Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.812645 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-stz5c" event={"ID":"ff0a141c-3800-4c84-b1d6-17ec909d5d27","Type":"ContainerDied","Data":"af4d438849daa5d4e132ae67e3659b425f69071f5b4262fbdd5a5f7837f9cc04"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.812701 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-stz5c" event={"ID":"ff0a141c-3800-4c84-b1d6-17ec909d5d27","Type":"ContainerStarted","Data":"1b1f044e4baad409b2fd46ee476b0abf1db6187b6885076a6f23f0d06edcaa21"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.821656 4839 generic.go:334] "Generic (PLEG): container finished" podID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerID="89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab" exitCode=0 Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.821765 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2p75" event={"ID":"b471c70a-15fe-4b5f-aa6f-354848925ace","Type":"ContainerDied","Data":"89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.821797 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2p75" event={"ID":"b471c70a-15fe-4b5f-aa6f-354848925ace","Type":"ContainerStarted","Data":"dd897366779f7f72927bd14fe45246f0517568162e089987e03c69a68e6e97ac"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.831573 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m6927" event={"ID":"ff369d3d-8b9b-4e20-9e1d-c404e3b45f1b","Type":"ContainerStarted","Data":"bb7f6e236a94550e780223055359c03d736404ad92506740d42468cfa31b9b40"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.835170 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" event={"ID":"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4","Type":"ContainerStarted","Data":"f107958b07275167df4df07296e659a3f5aa98f6d6df1393c397eceb9cfb2940"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.835216 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" event={"ID":"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4","Type":"ContainerStarted","Data":"a78f243efb241fbca05b215f8c65e6574111e14e5da972a71e5168a75a5b6878"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.843578 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" event={"ID":"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b","Type":"ContainerStarted","Data":"0177a0f83f670cb0702c9a49b2d5a89570ed2811de7ebd62322f15009e27943c"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.843627 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" event={"ID":"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b","Type":"ContainerStarted","Data":"21f721d6408dc116f69076bc5618e10c7c26a5c37b47b20c778fa0e8093cc253"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.844398 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.850309 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.850495 4839 generic.go:334] "Generic (PLEG): container finished" podID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerID="4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1" exitCode=0 Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.850552 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerDied","Data":"4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.850572 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerStarted","Data":"bc1ef7693019441d1a87decbad2ec0166bf4bf3d76d5293cf63f3e4acd0fd5cf"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.865738 4839 generic.go:334] "Generic (PLEG): container finished" podID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerID="2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c" exitCode=0 Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.866843 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5p9bn" event={"ID":"8cc72c87-b6d6-4384-ac6e-fb12af841f66","Type":"ContainerDied","Data":"2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.866874 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5p9bn" event={"ID":"8cc72c87-b6d6-4384-ac6e-fb12af841f66","Type":"ContainerStarted","Data":"34950ce8717a42cbebef568b404118724c4474152824aed30cea99a537383ff9"} Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.868285 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7fddl"] Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.870580 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f71939-65e0-4459-9c58-211f5cafd259" containerName="route-controller-manager" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.870723 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f71939-65e0-4459-9c58-211f5cafd259" containerName="route-controller-manager" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.870928 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f71939-65e0-4459-9c58-211f5cafd259" containerName="route-controller-manager" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.871932 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.874591 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.874958 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.374946409 +0000 UTC m=+216.019816144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.875347 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.888570 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fddl"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.903356 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-m6927" podStartSLOduration=170.903336364 podStartE2EDuration="2m50.903336364s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:33.900308864 +0000 UTC m=+215.545178609" watchObservedRunningTime="2026-02-27 19:37:33.903336364 +0000 UTC m=+215.548206099" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.955613 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.956433 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.959735 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.959836 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.965304 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.970180 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" podStartSLOduration=3.970154563 podStartE2EDuration="3.970154563s" podCreationTimestamp="2026-02-27 19:37:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:33.967887595 +0000 UTC m=+215.612757330" watchObservedRunningTime="2026-02-27 19:37:33.970154563 +0000 UTC m=+215.615024288" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.979382 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.984891 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.48485297 +0000 UTC m=+216.129722705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.986095 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w5zs\" (UniqueName: \"kubernetes.io/projected/67e667b9-304d-4052-8f0f-8d46d920aa7c-kube-api-access-8w5zs\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.986363 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-utilities\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.986426 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-catalog-content\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:33 crc kubenswrapper[4839]: I0227 19:37:33.986564 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:33 crc kubenswrapper[4839]: E0227 19:37:33.994711 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.494680263 +0000 UTC m=+216.139549998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.087874 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088190 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w5zs\" (UniqueName: \"kubernetes.io/projected/67e667b9-304d-4052-8f0f-8d46d920aa7c-kube-api-access-8w5zs\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.088321 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.58830149 +0000 UTC m=+216.233171225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088379 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-utilities\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088406 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-catalog-content\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088440 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088465 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088502 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.088895 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.588864927 +0000 UTC m=+216.233734662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.088963 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-utilities\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.089201 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-catalog-content\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.107338 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w5zs\" (UniqueName: \"kubernetes.io/projected/67e667b9-304d-4052-8f0f-8d46d920aa7c-kube-api-access-8w5zs\") pod \"redhat-marketplace-7fddl\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.189563 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.189735 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.190179 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.690161543 +0000 UTC m=+216.335031278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.190245 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.190272 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.190301 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.190558 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.190790 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.690782701 +0000 UTC m=+216.335652436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf6mj" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.213046 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.259468 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ff888"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.260536 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.285859 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff888"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.291658 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.292039 4839 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 19:37:34.792019495 +0000 UTC m=+216.436889230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.302504 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.346690 4839 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-27T19:37:33.509745547Z","Handler":null,"Name":""} Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.350338 4839 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.350369 4839 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.392954 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.393010 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-catalog-content\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.393036 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-utilities\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.393056 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n76bg\" (UniqueName: \"kubernetes.io/projected/68a8a402-5d27-4632-8cf2-aaccccfff95d-kube-api-access-n76bg\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.395485 4839 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.395518 4839 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.422424 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf6mj\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.494722 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.495102 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-catalog-content\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.495131 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-utilities\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.495153 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n76bg\" (UniqueName: \"kubernetes.io/projected/68a8a402-5d27-4632-8cf2-aaccccfff95d-kube-api-access-n76bg\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.496110 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-catalog-content\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.496197 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-utilities\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.510371 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.520173 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n76bg\" (UniqueName: \"kubernetes.io/projected/68a8a402-5d27-4632-8cf2-aaccccfff95d-kube-api-access-n76bg\") pod \"redhat-marketplace-ff888\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.560981 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.565859 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.566005 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5wrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d2p75_openshift-marketplace(b471c70a-15fe-4b5f-aa6f-354848925ace): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.567776 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.581885 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.582049 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8btrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-stz5c_openshift-marketplace(ff0a141c-3800-4c84-b1d6-17ec909d5d27): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.582826 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.583195 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.609895 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fddl"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.635505 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.651816 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.651978 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zm9fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgzcw_openshift-marketplace(0a4d0578-88dd-4754-81fd-ae28a709b698): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.653541 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.679228 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:34 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:34 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:34 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.679279 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.681118 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.681242 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvsn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5p9bn_openshift-marketplace(8cc72c87-b6d6-4384-ac6e-fb12af841f66): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.682372 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.830063 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff888"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.861458 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m6qmh"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.863002 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.865524 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 19:37:34 crc kubenswrapper[4839]: W0227 19:37:34.869352 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68a8a402_5d27_4632_8cf2_aaccccfff95d.slice/crio-2bfa82e174ea9fbabffe183a7d7d3dd3b62d4d47add2072e2b7a770b63adc1e1 WatchSource:0}: Error finding container 2bfa82e174ea9fbabffe183a7d7d3dd3b62d4d47add2072e2b7a770b63adc1e1: Status 404 returned error can't find the container with id 2bfa82e174ea9fbabffe183a7d7d3dd3b62d4d47add2072e2b7a770b63adc1e1 Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.871346 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m6qmh"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.910252 4839 generic.go:334] "Generic (PLEG): container finished" podID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerID="bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b" exitCode=0 Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.910550 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerDied","Data":"bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b"} Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.910575 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerStarted","Data":"d57f6811ea530e069dd3c348bfdcb5b9cf37d47bca3b133b5442cc82ecc62964"} Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.911098 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf6mj"] Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.923355 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63d060d1-ffbf-4eb9-b82f-f99752012ee7","Type":"ContainerStarted","Data":"c1dfa20be9367f49d18b08861346555bf9403102baefaf06d474ce708624cac4"} Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.938064 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" event={"ID":"1eb21c22-1c0f-4a33-93d8-d50cee6ecba4","Type":"ContainerStarted","Data":"63cdb05c6772b7833a28383e25e3232ccaf8464d2787e65dcf95ba495e42a427"} Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.950478 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.950783 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.950854 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:37:34 crc kubenswrapper[4839]: E0227 19:37:34.950892 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.986530 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.987389 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef6e180-79df-40ee-bf2b-96332e35c052" path="/var/lib/kubelet/pods/9ef6e180-79df-40ee-bf2b-96332e35c052/volumes" Feb 27 19:37:34 crc kubenswrapper[4839]: I0227 19:37:34.996827 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-txbn6" podStartSLOduration=12.996804036 podStartE2EDuration="12.996804036s" podCreationTimestamp="2026-02-27 19:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:34.961705231 +0000 UTC m=+216.606574976" watchObservedRunningTime="2026-02-27 19:37:34.996804036 +0000 UTC m=+216.641673781" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.010211 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-catalog-content\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.010278 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlf2r\" (UniqueName: \"kubernetes.io/projected/f11b8271-1de0-432d-8d78-2379862726a1-kube-api-access-nlf2r\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.010349 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-utilities\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.111785 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-utilities\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.111898 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-catalog-content\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.111944 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlf2r\" (UniqueName: \"kubernetes.io/projected/f11b8271-1de0-432d-8d78-2379862726a1-kube-api-access-nlf2r\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.112711 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-utilities\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.112920 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-catalog-content\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.136278 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlf2r\" (UniqueName: \"kubernetes.io/projected/f11b8271-1de0-432d-8d78-2379862726a1-kube-api-access-nlf2r\") pod \"redhat-operators-m6qmh\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.189096 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.214945 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.216088 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.216314 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.221404 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qfgcq" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.223560 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.254735 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.256113 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.258031 4839 patch_prober.go:28] interesting pod/console-f9d7485db-pc7k8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.258096 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pc7k8" podUID="e105188e-f3d1-4757-8134-0db7ce5cbbf2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.260132 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jn87v"] Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.261367 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.277938 4839 patch_prober.go:28] interesting pod/downloads-7954f5f757-82v8f container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.277976 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-82v8f" podUID="a4baae9f-93ba-40bf-9c28-01f13cdcaf43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.278471 4839 patch_prober.go:28] interesting pod/downloads-7954f5f757-82v8f container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.278493 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-82v8f" podUID="a4baae9f-93ba-40bf-9c28-01f13cdcaf43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.308228 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jn87v"] Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.360300 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6"] Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.361072 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.371240 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.371511 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.371683 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.372175 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.372288 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.372431 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.378166 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6"] Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.423517 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-utilities\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.423616 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9svsj\" (UniqueName: \"kubernetes.io/projected/6321a495-e9eb-4ca0-8adf-af955a1a87c0-kube-api-access-9svsj\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.423711 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-catalog-content\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.486237 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m6qmh"] Feb 27 19:37:35 crc kubenswrapper[4839]: W0227 19:37:35.492800 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf11b8271_1de0_432d_8d78_2379862726a1.slice/crio-1677b5828c5ec3c5f293826a708719dbfce76bc862659d0abe096d491698c334 WatchSource:0}: Error finding container 1677b5828c5ec3c5f293826a708719dbfce76bc862659d0abe096d491698c334: Status 404 returned error can't find the container with id 1677b5828c5ec3c5f293826a708719dbfce76bc862659d0abe096d491698c334 Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.527298 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnrvd\" (UniqueName: \"kubernetes.io/projected/0984e4ce-89b5-42e6-a11a-103080dd76de-kube-api-access-gnrvd\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.527448 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9svsj\" (UniqueName: \"kubernetes.io/projected/6321a495-e9eb-4ca0-8adf-af955a1a87c0-kube-api-access-9svsj\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.527852 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0984e4ce-89b5-42e6-a11a-103080dd76de-serving-cert\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.527893 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-client-ca\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.527944 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-catalog-content\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.527990 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-utilities\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.528013 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-config\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.528378 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-catalog-content\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.528623 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-utilities\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.548991 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9svsj\" (UniqueName: \"kubernetes.io/projected/6321a495-e9eb-4ca0-8adf-af955a1a87c0-kube-api-access-9svsj\") pod \"redhat-operators-jn87v\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.586418 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.629349 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0984e4ce-89b5-42e6-a11a-103080dd76de-serving-cert\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.629391 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-client-ca\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.629434 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-config\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.629468 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnrvd\" (UniqueName: \"kubernetes.io/projected/0984e4ce-89b5-42e6-a11a-103080dd76de-kube-api-access-gnrvd\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.630911 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-client-ca\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.632287 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-config\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.632959 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0984e4ce-89b5-42e6-a11a-103080dd76de-serving-cert\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: E0227 19:37:35.634843 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:37:35 crc kubenswrapper[4839]: E0227 19:37:35.634978 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w5zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7fddl_openshift-marketplace(67e667b9-304d-4052-8f0f-8d46d920aa7c): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:35 crc kubenswrapper[4839]: E0227 19:37:35.636099 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.655715 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnrvd\" (UniqueName: \"kubernetes.io/projected/0984e4ce-89b5-42e6-a11a-103080dd76de-kube-api-access-gnrvd\") pod \"route-controller-manager-5bb74cc8fd-5rxb6\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.677220 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.679387 4839 patch_prober.go:28] interesting pod/router-default-5444994796-85jfr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 19:37:35 crc kubenswrapper[4839]: [-]has-synced failed: reason withheld Feb 27 19:37:35 crc kubenswrapper[4839]: [+]process-running ok Feb 27 19:37:35 crc kubenswrapper[4839]: healthz check failed Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.679447 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85jfr" podUID="81d8a2c8-0ce3-4a9e-8d41-370b2107a840" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.686405 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.775812 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jn87v"] Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.884958 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6"] Feb 27 19:37:35 crc kubenswrapper[4839]: W0227 19:37:35.888946 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0984e4ce_89b5_42e6_a11a_103080dd76de.slice/crio-91f2e45af178fdc05984d07a8d7056de7251e54360667b26a37dead83c9f796a WatchSource:0}: Error finding container 91f2e45af178fdc05984d07a8d7056de7251e54360667b26a37dead83c9f796a: Status 404 returned error can't find the container with id 91f2e45af178fdc05984d07a8d7056de7251e54360667b26a37dead83c9f796a Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.952986 4839 generic.go:334] "Generic (PLEG): container finished" podID="f11b8271-1de0-432d-8d78-2379862726a1" containerID="6f6bc15dd0710573d6c72e4b110fae722d4d86042550b9431787d9bc0a0e9d34" exitCode=0 Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.953064 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6qmh" event={"ID":"f11b8271-1de0-432d-8d78-2379862726a1","Type":"ContainerDied","Data":"6f6bc15dd0710573d6c72e4b110fae722d4d86042550b9431787d9bc0a0e9d34"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.953093 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6qmh" event={"ID":"f11b8271-1de0-432d-8d78-2379862726a1","Type":"ContainerStarted","Data":"1677b5828c5ec3c5f293826a708719dbfce76bc862659d0abe096d491698c334"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.955722 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" event={"ID":"0984e4ce-89b5-42e6-a11a-103080dd76de","Type":"ContainerStarted","Data":"91f2e45af178fdc05984d07a8d7056de7251e54360667b26a37dead83c9f796a"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.959288 4839 generic.go:334] "Generic (PLEG): container finished" podID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerID="f5e10603dc3d7f8858cdaf045d7be6d1b373e231e3f6d783d7341eb4d6266e33" exitCode=0 Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.959343 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerDied","Data":"f5e10603dc3d7f8858cdaf045d7be6d1b373e231e3f6d783d7341eb4d6266e33"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.959364 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerStarted","Data":"50e30c9a5fd7cf82eb6f29728385ee1c045c45bc11efb9f14e239ab2547c14ec"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.961769 4839 generic.go:334] "Generic (PLEG): container finished" podID="68a8a402-5d27-4632-8cf2-aaccccfff95d" containerID="f8176690a7343e0e78625f76fa3e014e129765fac6f6dac057970c1bced3a2a6" exitCode=0 Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.961844 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff888" event={"ID":"68a8a402-5d27-4632-8cf2-aaccccfff95d","Type":"ContainerDied","Data":"f8176690a7343e0e78625f76fa3e014e129765fac6f6dac057970c1bced3a2a6"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.961873 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff888" event={"ID":"68a8a402-5d27-4632-8cf2-aaccccfff95d","Type":"ContainerStarted","Data":"2bfa82e174ea9fbabffe183a7d7d3dd3b62d4d47add2072e2b7a770b63adc1e1"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.965210 4839 generic.go:334] "Generic (PLEG): container finished" podID="63d060d1-ffbf-4eb9-b82f-f99752012ee7" containerID="3a574bceb1cf378efdf4edf07c5f1e0c2bec8a1a5e6ba1a73443ce81f6140294" exitCode=0 Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.965356 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63d060d1-ffbf-4eb9-b82f-f99752012ee7","Type":"ContainerDied","Data":"3a574bceb1cf378efdf4edf07c5f1e0c2bec8a1a5e6ba1a73443ce81f6140294"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.966659 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" event={"ID":"e4bdb150-3f14-4eac-9f05-e10a515c7cb4","Type":"ContainerStarted","Data":"533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.966900 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.966917 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" event={"ID":"e4bdb150-3f14-4eac-9f05-e10a515c7cb4","Type":"ContainerStarted","Data":"6aaa1982c32bed4aa521079631d6e92aacc86204cdce13b3f400e645bc76e15a"} Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.971080 4839 generic.go:334] "Generic (PLEG): container finished" podID="4b927691-5423-4670-8443-c0a6a9d3bc13" containerID="12ad453b448f74de4d775ae89cccfc7a64b1bb9268c9e9e368e2e7b88186e9a1" exitCode=0 Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.971224 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" event={"ID":"4b927691-5423-4670-8443-c0a6a9d3bc13","Type":"ContainerDied","Data":"12ad453b448f74de4d775ae89cccfc7a64b1bb9268c9e9e368e2e7b88186e9a1"} Feb 27 19:37:35 crc kubenswrapper[4839]: E0227 19:37:35.978088 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.979193 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xm448" Feb 27 19:37:35 crc kubenswrapper[4839]: I0227 19:37:35.989680 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" podStartSLOduration=172.989647103 podStartE2EDuration="2m52.989647103s" podCreationTimestamp="2026-02-27 19:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:35.988895371 +0000 UTC m=+217.633765126" watchObservedRunningTime="2026-02-27 19:37:35.989647103 +0000 UTC m=+217.634516838" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.581977 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.582327 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nlf2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-m6qmh_openshift-marketplace(f11b8271-1de0-432d-8d78-2379862726a1): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.583518 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-m6qmh" podUID="f11b8271-1de0-432d-8d78-2379862726a1" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.591980 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.592231 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n76bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ff888_openshift-marketplace(68a8a402-5d27-4632-8cf2-aaccccfff95d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.593463 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:37:36 crc kubenswrapper[4839]: I0227 19:37:36.687228 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:36 crc kubenswrapper[4839]: I0227 19:37:36.689279 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-85jfr" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.708469 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.708687 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9svsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jn87v_openshift-marketplace(6321a495-e9eb-4ca0-8adf-af955a1a87c0): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.709833 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-jn87v" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" Feb 27 19:37:36 crc kubenswrapper[4839]: I0227 19:37:36.976845 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" event={"ID":"0984e4ce-89b5-42e6-a11a-103080dd76de","Type":"ContainerStarted","Data":"b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11"} Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.983891 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.984023 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jn87v" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" Feb 27 19:37:36 crc kubenswrapper[4839]: E0227 19:37:36.985428 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-m6qmh" podUID="f11b8271-1de0-432d-8d78-2379862726a1" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.045977 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" podStartSLOduration=7.045957071 podStartE2EDuration="7.045957071s" podCreationTimestamp="2026-02-27 19:37:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:37.043788286 +0000 UTC m=+218.688658031" watchObservedRunningTime="2026-02-27 19:37:37.045957071 +0000 UTC m=+218.690826806" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.286266 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.291253 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.427742 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 19:37:37 crc kubenswrapper[4839]: E0227 19:37:37.427940 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d060d1-ffbf-4eb9-b82f-f99752012ee7" containerName="pruner" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.427951 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d060d1-ffbf-4eb9-b82f-f99752012ee7" containerName="pruner" Feb 27 19:37:37 crc kubenswrapper[4839]: E0227 19:37:37.427959 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b927691-5423-4670-8443-c0a6a9d3bc13" containerName="collect-profiles" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.427965 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b927691-5423-4670-8443-c0a6a9d3bc13" containerName="collect-profiles" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.428053 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b927691-5423-4670-8443-c0a6a9d3bc13" containerName="collect-profiles" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.428065 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d060d1-ffbf-4eb9-b82f-f99752012ee7" containerName="pruner" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.428398 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.431967 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.432851 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.440975 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455521 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jdxz\" (UniqueName: \"kubernetes.io/projected/4b927691-5423-4670-8443-c0a6a9d3bc13-kube-api-access-2jdxz\") pod \"4b927691-5423-4670-8443-c0a6a9d3bc13\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455582 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kubelet-dir\") pod \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455645 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b927691-5423-4670-8443-c0a6a9d3bc13-config-volume\") pod \"4b927691-5423-4670-8443-c0a6a9d3bc13\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455684 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b927691-5423-4670-8443-c0a6a9d3bc13-secret-volume\") pod \"4b927691-5423-4670-8443-c0a6a9d3bc13\" (UID: \"4b927691-5423-4670-8443-c0a6a9d3bc13\") " Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455725 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kube-api-access\") pod \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\" (UID: \"63d060d1-ffbf-4eb9-b82f-f99752012ee7\") " Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455736 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "63d060d1-ffbf-4eb9-b82f-f99752012ee7" (UID: "63d060d1-ffbf-4eb9-b82f-f99752012ee7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.455959 4839 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.456471 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b927691-5423-4670-8443-c0a6a9d3bc13-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b927691-5423-4670-8443-c0a6a9d3bc13" (UID: "4b927691-5423-4670-8443-c0a6a9d3bc13"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.463913 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "63d060d1-ffbf-4eb9-b82f-f99752012ee7" (UID: "63d060d1-ffbf-4eb9-b82f-f99752012ee7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.464194 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b927691-5423-4670-8443-c0a6a9d3bc13-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b927691-5423-4670-8443-c0a6a9d3bc13" (UID: "4b927691-5423-4670-8443-c0a6a9d3bc13"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.464346 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b927691-5423-4670-8443-c0a6a9d3bc13-kube-api-access-2jdxz" (OuterVolumeSpecName: "kube-api-access-2jdxz") pod "4b927691-5423-4670-8443-c0a6a9d3bc13" (UID: "4b927691-5423-4670-8443-c0a6a9d3bc13"). InnerVolumeSpecName "kube-api-access-2jdxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.557265 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/878fd70c-2b2a-4916-83cb-31f96f3216cd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.557420 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/878fd70c-2b2a-4916-83cb-31f96f3216cd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.557880 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jdxz\" (UniqueName: \"kubernetes.io/projected/4b927691-5423-4670-8443-c0a6a9d3bc13-kube-api-access-2jdxz\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.557940 4839 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b927691-5423-4670-8443-c0a6a9d3bc13-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.557958 4839 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b927691-5423-4670-8443-c0a6a9d3bc13-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.557976 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63d060d1-ffbf-4eb9-b82f-f99752012ee7-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.658781 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/878fd70c-2b2a-4916-83cb-31f96f3216cd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.658909 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/878fd70c-2b2a-4916-83cb-31f96f3216cd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.658914 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/878fd70c-2b2a-4916-83cb-31f96f3216cd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.677687 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/878fd70c-2b2a-4916-83cb-31f96f3216cd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.742429 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.926584 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 19:37:37 crc kubenswrapper[4839]: W0227 19:37:37.938324 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod878fd70c_2b2a_4916_83cb_31f96f3216cd.slice/crio-86cf3b784515c3f9d634f13c57e262f60074c96f76977ec85a98ca4aebdd4eb3 WatchSource:0}: Error finding container 86cf3b784515c3f9d634f13c57e262f60074c96f76977ec85a98ca4aebdd4eb3: Status 404 returned error can't find the container with id 86cf3b784515c3f9d634f13c57e262f60074c96f76977ec85a98ca4aebdd4eb3 Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.987848 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"878fd70c-2b2a-4916-83cb-31f96f3216cd","Type":"ContainerStarted","Data":"86cf3b784515c3f9d634f13c57e262f60074c96f76977ec85a98ca4aebdd4eb3"} Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.989792 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.993053 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63d060d1-ffbf-4eb9-b82f-f99752012ee7","Type":"ContainerDied","Data":"c1dfa20be9367f49d18b08861346555bf9403102baefaf06d474ce708624cac4"} Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.993091 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1dfa20be9367f49d18b08861346555bf9403102baefaf06d474ce708624cac4" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.995489 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" event={"ID":"4b927691-5423-4670-8443-c0a6a9d3bc13","Type":"ContainerDied","Data":"8b8af4a16b1114faf54cee5113c79f1e5ee6c2877c6f71413fade4419f286dac"} Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.995567 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b8af4a16b1114faf54cee5113c79f1e5ee6c2877c6f71413fade4419f286dac" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.995830 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537010-7cdv5" Feb 27 19:37:37 crc kubenswrapper[4839]: I0227 19:37:37.996217 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:38 crc kubenswrapper[4839]: I0227 19:37:38.008242 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:38 crc kubenswrapper[4839]: I0227 19:37:38.083552 4839 ???:1] "http: TLS handshake error from 192.168.126.11:37342: no serving certificate available for the kubelet" Feb 27 19:37:38 crc kubenswrapper[4839]: I0227 19:37:38.723889 4839 ???:1] "http: TLS handshake error from 192.168.126.11:37356: no serving certificate available for the kubelet" Feb 27 19:37:39 crc kubenswrapper[4839]: I0227 19:37:39.001225 4839 generic.go:334] "Generic (PLEG): container finished" podID="878fd70c-2b2a-4916-83cb-31f96f3216cd" containerID="254c02cc56cfb39b054b6676be9593bbeed57d1aee1987539be2fbcdaa36a01d" exitCode=0 Feb 27 19:37:39 crc kubenswrapper[4839]: I0227 19:37:39.001443 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"878fd70c-2b2a-4916-83cb-31f96f3216cd","Type":"ContainerDied","Data":"254c02cc56cfb39b054b6676be9593bbeed57d1aee1987539be2fbcdaa36a01d"} Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.280261 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.397778 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/878fd70c-2b2a-4916-83cb-31f96f3216cd-kubelet-dir\") pod \"878fd70c-2b2a-4916-83cb-31f96f3216cd\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.397910 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/878fd70c-2b2a-4916-83cb-31f96f3216cd-kube-api-access\") pod \"878fd70c-2b2a-4916-83cb-31f96f3216cd\" (UID: \"878fd70c-2b2a-4916-83cb-31f96f3216cd\") " Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.397936 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/878fd70c-2b2a-4916-83cb-31f96f3216cd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "878fd70c-2b2a-4916-83cb-31f96f3216cd" (UID: "878fd70c-2b2a-4916-83cb-31f96f3216cd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.398201 4839 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/878fd70c-2b2a-4916-83cb-31f96f3216cd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.402392 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/878fd70c-2b2a-4916-83cb-31f96f3216cd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "878fd70c-2b2a-4916-83cb-31f96f3216cd" (UID: "878fd70c-2b2a-4916-83cb-31f96f3216cd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:40 crc kubenswrapper[4839]: I0227 19:37:40.499399 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/878fd70c-2b2a-4916-83cb-31f96f3216cd-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:41 crc kubenswrapper[4839]: I0227 19:37:41.010895 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"878fd70c-2b2a-4916-83cb-31f96f3216cd","Type":"ContainerDied","Data":"86cf3b784515c3f9d634f13c57e262f60074c96f76977ec85a98ca4aebdd4eb3"} Feb 27 19:37:41 crc kubenswrapper[4839]: I0227 19:37:41.011174 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86cf3b784515c3f9d634f13c57e262f60074c96f76977ec85a98ca4aebdd4eb3" Feb 27 19:37:41 crc kubenswrapper[4839]: I0227 19:37:41.010955 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 19:37:41 crc kubenswrapper[4839]: I0227 19:37:41.088433 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-t69c7" Feb 27 19:37:44 crc kubenswrapper[4839]: E0227 19:37:44.005211 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:37:44 crc kubenswrapper[4839]: E0227 19:37:44.005654 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:37:44 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:37:44 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmg87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537016-rt648_openshift-infra(95e89fb2-3271-46bd-878b-6ff642b17214): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:37:44 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:37:44 crc kubenswrapper[4839]: E0227 19:37:44.006904 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:37:45 crc kubenswrapper[4839]: I0227 19:37:45.261320 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:45 crc kubenswrapper[4839]: I0227 19:37:45.265685 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pc7k8" Feb 27 19:37:45 crc kubenswrapper[4839]: I0227 19:37:45.291027 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-82v8f" Feb 27 19:37:47 crc kubenswrapper[4839]: E0227 19:37:47.501933 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:37:47 crc kubenswrapper[4839]: E0227 19:37:47.502431 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zm9fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgzcw_openshift-marketplace(0a4d0578-88dd-4754-81fd-ae28a709b698): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:47 crc kubenswrapper[4839]: E0227 19:37:47.503633 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:37:47 crc kubenswrapper[4839]: E0227 19:37:47.882833 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 19:37:47 crc kubenswrapper[4839]: E0227 19:37:47.883554 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8btrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-stz5c_openshift-marketplace(ff0a141c-3800-4c84-b1d6-17ec909d5d27): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:47 crc kubenswrapper[4839]: E0227 19:37:47.884886 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:37:48 crc kubenswrapper[4839]: E0227 19:37:48.720317 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:37:48 crc kubenswrapper[4839]: E0227 19:37:48.720497 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w5zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7fddl_openshift-marketplace(67e667b9-304d-4052-8f0f-8d46d920aa7c): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:48 crc kubenswrapper[4839]: E0227 19:37:48.720914 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:37:48 crc kubenswrapper[4839]: E0227 19:37:48.720990 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5wrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d2p75_openshift-marketplace(b471c70a-15fe-4b5f-aa6f-354848925ace): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:48 crc kubenswrapper[4839]: E0227 19:37:48.721973 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:37:48 crc kubenswrapper[4839]: E0227 19:37:48.722051 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:37:49 crc kubenswrapper[4839]: I0227 19:37:49.298783 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs"] Feb 27 19:37:49 crc kubenswrapper[4839]: I0227 19:37:49.299126 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" podUID="0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" containerName="controller-manager" containerID="cri-o://0177a0f83f670cb0702c9a49b2d5a89570ed2811de7ebd62322f15009e27943c" gracePeriod=30 Feb 27 19:37:49 crc kubenswrapper[4839]: I0227 19:37:49.335311 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6"] Feb 27 19:37:49 crc kubenswrapper[4839]: I0227 19:37:49.335540 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" podUID="0984e4ce-89b5-42e6-a11a-103080dd76de" containerName="route-controller-manager" containerID="cri-o://b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11" gracePeriod=30 Feb 27 19:37:49 crc kubenswrapper[4839]: E0227 19:37:49.641820 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 19:37:49 crc kubenswrapper[4839]: E0227 19:37:49.641988 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvsn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5p9bn_openshift-marketplace(8cc72c87-b6d6-4384-ac6e-fb12af841f66): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:49 crc kubenswrapper[4839]: E0227 19:37:49.643991 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:37:50 crc kubenswrapper[4839]: I0227 19:37:50.058856 4839 generic.go:334] "Generic (PLEG): container finished" podID="0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" containerID="0177a0f83f670cb0702c9a49b2d5a89570ed2811de7ebd62322f15009e27943c" exitCode=0 Feb 27 19:37:50 crc kubenswrapper[4839]: I0227 19:37:50.058897 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" event={"ID":"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b","Type":"ContainerDied","Data":"0177a0f83f670cb0702c9a49b2d5a89570ed2811de7ebd62322f15009e27943c"} Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.231513 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.231963 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n76bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ff888_openshift-marketplace(68a8a402-5d27-4632-8cf2-aaccccfff95d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.233509 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.265109 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.265260 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9svsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jn87v_openshift-marketplace(6321a495-e9eb-4ca0-8adf-af955a1a87c0): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.266474 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-jn87v" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.591069 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.621225 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5967b47cf9-jvh6q"] Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.621633 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="878fd70c-2b2a-4916-83cb-31f96f3216cd" containerName="pruner" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.621682 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="878fd70c-2b2a-4916-83cb-31f96f3216cd" containerName="pruner" Feb 27 19:37:51 crc kubenswrapper[4839]: E0227 19:37:51.621710 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" containerName="controller-manager" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.621719 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" containerName="controller-manager" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.621850 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="878fd70c-2b2a-4916-83cb-31f96f3216cd" containerName="pruner" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.621869 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" containerName="controller-manager" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.622353 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.626860 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5967b47cf9-jvh6q"] Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740154 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-client-ca\") pod \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740253 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-config\") pod \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740335 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-serving-cert\") pod \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740455 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xx8n\" (UniqueName: \"kubernetes.io/projected/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-kube-api-access-2xx8n\") pod \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740492 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-proxy-ca-bundles\") pod \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\" (UID: \"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b\") " Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740824 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-proxy-ca-bundles\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740909 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb84b0aa-f8fe-4074-aa11-13d147f52d52-serving-cert\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.740987 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-client-ca\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.741200 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4rdq\" (UniqueName: \"kubernetes.io/projected/eb84b0aa-f8fe-4074-aa11-13d147f52d52-kube-api-access-k4rdq\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.741317 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-config\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.741563 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" (UID: "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.741600 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-client-ca" (OuterVolumeSpecName: "client-ca") pod "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" (UID: "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.741797 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-config" (OuterVolumeSpecName: "config") pod "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" (UID: "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.746901 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" (UID: "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.748149 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-kube-api-access-2xx8n" (OuterVolumeSpecName: "kube-api-access-2xx8n") pod "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" (UID: "0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b"). InnerVolumeSpecName "kube-api-access-2xx8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.842509 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4rdq\" (UniqueName: \"kubernetes.io/projected/eb84b0aa-f8fe-4074-aa11-13d147f52d52-kube-api-access-k4rdq\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.842986 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-config\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843044 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-proxy-ca-bundles\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843086 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb84b0aa-f8fe-4074-aa11-13d147f52d52-serving-cert\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843140 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-client-ca\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843200 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xx8n\" (UniqueName: \"kubernetes.io/projected/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-kube-api-access-2xx8n\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843220 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843240 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843257 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.843275 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.844274 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-proxy-ca-bundles\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.844416 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-config\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.844516 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-client-ca\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.848366 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb84b0aa-f8fe-4074-aa11-13d147f52d52-serving-cert\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:51 crc kubenswrapper[4839]: I0227 19:37:51.957202 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.046900 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnrvd\" (UniqueName: \"kubernetes.io/projected/0984e4ce-89b5-42e6-a11a-103080dd76de-kube-api-access-gnrvd\") pod \"0984e4ce-89b5-42e6-a11a-103080dd76de\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.046977 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-client-ca\") pod \"0984e4ce-89b5-42e6-a11a-103080dd76de\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.047074 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0984e4ce-89b5-42e6-a11a-103080dd76de-serving-cert\") pod \"0984e4ce-89b5-42e6-a11a-103080dd76de\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.047120 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-config\") pod \"0984e4ce-89b5-42e6-a11a-103080dd76de\" (UID: \"0984e4ce-89b5-42e6-a11a-103080dd76de\") " Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.048420 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-config" (OuterVolumeSpecName: "config") pod "0984e4ce-89b5-42e6-a11a-103080dd76de" (UID: "0984e4ce-89b5-42e6-a11a-103080dd76de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.048553 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-client-ca" (OuterVolumeSpecName: "client-ca") pod "0984e4ce-89b5-42e6-a11a-103080dd76de" (UID: "0984e4ce-89b5-42e6-a11a-103080dd76de"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.054532 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0984e4ce-89b5-42e6-a11a-103080dd76de-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0984e4ce-89b5-42e6-a11a-103080dd76de" (UID: "0984e4ce-89b5-42e6-a11a-103080dd76de"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.054711 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0984e4ce-89b5-42e6-a11a-103080dd76de-kube-api-access-gnrvd" (OuterVolumeSpecName: "kube-api-access-gnrvd") pod "0984e4ce-89b5-42e6-a11a-103080dd76de" (UID: "0984e4ce-89b5-42e6-a11a-103080dd76de"). InnerVolumeSpecName "kube-api-access-gnrvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.055326 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4rdq\" (UniqueName: \"kubernetes.io/projected/eb84b0aa-f8fe-4074-aa11-13d147f52d52-kube-api-access-k4rdq\") pod \"controller-manager-5967b47cf9-jvh6q\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.073579 4839 generic.go:334] "Generic (PLEG): container finished" podID="0984e4ce-89b5-42e6-a11a-103080dd76de" containerID="b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11" exitCode=0 Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.073624 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.073677 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" event={"ID":"0984e4ce-89b5-42e6-a11a-103080dd76de","Type":"ContainerDied","Data":"b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11"} Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.073714 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6" event={"ID":"0984e4ce-89b5-42e6-a11a-103080dd76de","Type":"ContainerDied","Data":"91f2e45af178fdc05984d07a8d7056de7251e54360667b26a37dead83c9f796a"} Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.073729 4839 scope.go:117] "RemoveContainer" containerID="b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.075006 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" event={"ID":"0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b","Type":"ContainerDied","Data":"21f721d6408dc116f69076bc5618e10c7c26a5c37b47b20c778fa0e8093cc253"} Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.075079 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.107516 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6"] Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.107570 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bb74cc8fd-5rxb6"] Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.110414 4839 scope.go:117] "RemoveContainer" containerID="b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11" Feb 27 19:37:52 crc kubenswrapper[4839]: E0227 19:37:52.111389 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11\": container with ID starting with b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11 not found: ID does not exist" containerID="b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.111418 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11"} err="failed to get container status \"b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11\": rpc error: code = NotFound desc = could not find container \"b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11\": container with ID starting with b9f93dac965555aa598e2010f3385827ddcc7b5b206b02f5694c8cf4f5206a11 not found: ID does not exist" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.111439 4839 scope.go:117] "RemoveContainer" containerID="0177a0f83f670cb0702c9a49b2d5a89570ed2811de7ebd62322f15009e27943c" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.118162 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs"] Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.121369 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-9cdb8d79d-zhqrs"] Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.148320 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.148347 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnrvd\" (UniqueName: \"kubernetes.io/projected/0984e4ce-89b5-42e6-a11a-103080dd76de-kube-api-access-gnrvd\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.148357 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0984e4ce-89b5-42e6-a11a-103080dd76de-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.148368 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0984e4ce-89b5-42e6-a11a-103080dd76de-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.244053 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.420208 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5967b47cf9-jvh6q"] Feb 27 19:37:52 crc kubenswrapper[4839]: W0227 19:37:52.428463 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb84b0aa_f8fe_4074_aa11_13d147f52d52.slice/crio-526af1c9e54c79ff3d15e1ef95bb057fbcc9b5b65ed627d6ab5062a22644cc98 WatchSource:0}: Error finding container 526af1c9e54c79ff3d15e1ef95bb057fbcc9b5b65ed627d6ab5062a22644cc98: Status 404 returned error can't find the container with id 526af1c9e54c79ff3d15e1ef95bb057fbcc9b5b65ed627d6ab5062a22644cc98 Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.682294 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.682349 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.972443 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0984e4ce-89b5-42e6-a11a-103080dd76de" path="/var/lib/kubelet/pods/0984e4ce-89b5-42e6-a11a-103080dd76de/volumes" Feb 27 19:37:52 crc kubenswrapper[4839]: I0227 19:37:52.973608 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b" path="/var/lib/kubelet/pods/0cd6064c-bb73-4d5a-a6b0-b2ad66c5ee2b/volumes" Feb 27 19:37:53 crc kubenswrapper[4839]: E0227 19:37:53.066378 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:37:53 crc kubenswrapper[4839]: E0227 19:37:53.066537 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nlf2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-m6qmh_openshift-marketplace(f11b8271-1de0-432d-8d78-2379862726a1): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:37:53 crc kubenswrapper[4839]: E0227 19:37:53.068081 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-m6qmh" podUID="f11b8271-1de0-432d-8d78-2379862726a1" Feb 27 19:37:53 crc kubenswrapper[4839]: I0227 19:37:53.083844 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" event={"ID":"eb84b0aa-f8fe-4074-aa11-13d147f52d52","Type":"ContainerStarted","Data":"9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97"} Feb 27 19:37:53 crc kubenswrapper[4839]: I0227 19:37:53.083886 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" event={"ID":"eb84b0aa-f8fe-4074-aa11-13d147f52d52","Type":"ContainerStarted","Data":"526af1c9e54c79ff3d15e1ef95bb057fbcc9b5b65ed627d6ab5062a22644cc98"} Feb 27 19:37:53 crc kubenswrapper[4839]: I0227 19:37:53.084204 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:53 crc kubenswrapper[4839]: I0227 19:37:53.087679 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:37:53 crc kubenswrapper[4839]: I0227 19:37:53.101230 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" podStartSLOduration=4.101214083 podStartE2EDuration="4.101214083s" podCreationTimestamp="2026-02-27 19:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:53.097854593 +0000 UTC m=+234.742724318" watchObservedRunningTime="2026-02-27 19:37:53.101214083 +0000 UTC m=+234.746083818" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.371600 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk"] Feb 27 19:37:54 crc kubenswrapper[4839]: E0227 19:37:54.371851 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0984e4ce-89b5-42e6-a11a-103080dd76de" containerName="route-controller-manager" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.371866 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="0984e4ce-89b5-42e6-a11a-103080dd76de" containerName="route-controller-manager" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.371981 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="0984e4ce-89b5-42e6-a11a-103080dd76de" containerName="route-controller-manager" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.372407 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.375763 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.376390 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.377596 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.377795 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.378075 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.378735 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.381980 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk"] Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.475948 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/038a5536-0e8e-478d-b743-be2c50c910fd-serving-cert\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.476010 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr588\" (UniqueName: \"kubernetes.io/projected/038a5536-0e8e-478d-b743-be2c50c910fd-kube-api-access-zr588\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.476064 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-config\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.476162 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-client-ca\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.577414 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/038a5536-0e8e-478d-b743-be2c50c910fd-serving-cert\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.577468 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr588\" (UniqueName: \"kubernetes.io/projected/038a5536-0e8e-478d-b743-be2c50c910fd-kube-api-access-zr588\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.577521 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-config\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.577591 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-client-ca\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.578424 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-client-ca\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.578592 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-config\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.583088 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/038a5536-0e8e-478d-b743-be2c50c910fd-serving-cert\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.593387 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr588\" (UniqueName: \"kubernetes.io/projected/038a5536-0e8e-478d-b743-be2c50c910fd-kube-api-access-zr588\") pod \"route-controller-manager-8449df6b88-c2wnk\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.642059 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:37:54 crc kubenswrapper[4839]: I0227 19:37:54.701044 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:55 crc kubenswrapper[4839]: I0227 19:37:55.121218 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk"] Feb 27 19:37:55 crc kubenswrapper[4839]: W0227 19:37:55.132039 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod038a5536_0e8e_478d_b743_be2c50c910fd.slice/crio-0c0d368bb9cc30043d7832f8d16d2a7089f8d1c5581e99063279140d58b52def WatchSource:0}: Error finding container 0c0d368bb9cc30043d7832f8d16d2a7089f8d1c5581e99063279140d58b52def: Status 404 returned error can't find the container with id 0c0d368bb9cc30043d7832f8d16d2a7089f8d1c5581e99063279140d58b52def Feb 27 19:37:56 crc kubenswrapper[4839]: I0227 19:37:56.102826 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" event={"ID":"038a5536-0e8e-478d-b743-be2c50c910fd","Type":"ContainerStarted","Data":"b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310"} Feb 27 19:37:56 crc kubenswrapper[4839]: I0227 19:37:56.103323 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" event={"ID":"038a5536-0e8e-478d-b743-be2c50c910fd","Type":"ContainerStarted","Data":"0c0d368bb9cc30043d7832f8d16d2a7089f8d1c5581e99063279140d58b52def"} Feb 27 19:37:56 crc kubenswrapper[4839]: I0227 19:37:56.103349 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:56 crc kubenswrapper[4839]: I0227 19:37:56.108802 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:37:56 crc kubenswrapper[4839]: I0227 19:37:56.129521 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" podStartSLOduration=7.129504077 podStartE2EDuration="7.129504077s" podCreationTimestamp="2026-02-27 19:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:37:56.12154434 +0000 UTC m=+237.766414115" watchObservedRunningTime="2026-02-27 19:37:56.129504077 +0000 UTC m=+237.774373812" Feb 27 19:37:56 crc kubenswrapper[4839]: E0227 19:37:56.968510 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:37:59 crc kubenswrapper[4839]: I0227 19:37:59.225126 4839 ???:1] "http: TLS handshake error from 192.168.126.11:54982: no serving certificate available for the kubelet" Feb 27 19:37:59 crc kubenswrapper[4839]: E0227 19:37:59.966497 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:37:59 crc kubenswrapper[4839]: E0227 19:37:59.966514 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.130824 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537018-dbqzs"] Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.131567 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.133453 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44lp8" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.136550 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537018-dbqzs"] Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.254934 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j52c5\" (UniqueName: \"kubernetes.io/projected/499a12e9-ae64-4d99-93ed-e6af775581c0-kube-api-access-j52c5\") pod \"auto-csr-approver-29537018-dbqzs\" (UID: \"499a12e9-ae64-4d99-93ed-e6af775581c0\") " pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.356541 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j52c5\" (UniqueName: \"kubernetes.io/projected/499a12e9-ae64-4d99-93ed-e6af775581c0-kube-api-access-j52c5\") pod \"auto-csr-approver-29537018-dbqzs\" (UID: \"499a12e9-ae64-4d99-93ed-e6af775581c0\") " pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.375801 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j52c5\" (UniqueName: \"kubernetes.io/projected/499a12e9-ae64-4d99-93ed-e6af775581c0-kube-api-access-j52c5\") pod \"auto-csr-approver-29537018-dbqzs\" (UID: \"499a12e9-ae64-4d99-93ed-e6af775581c0\") " pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.446775 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:00 crc kubenswrapper[4839]: I0227 19:38:00.831057 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537018-dbqzs"] Feb 27 19:38:00 crc kubenswrapper[4839]: W0227 19:38:00.835994 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod499a12e9_ae64_4d99_93ed_e6af775581c0.slice/crio-ac883ad2b4650d4ed67d251db2ca2bc803ab94aad7d17136ba7efb16b6b67ddd WatchSource:0}: Error finding container ac883ad2b4650d4ed67d251db2ca2bc803ab94aad7d17136ba7efb16b6b67ddd: Status 404 returned error can't find the container with id ac883ad2b4650d4ed67d251db2ca2bc803ab94aad7d17136ba7efb16b6b67ddd Feb 27 19:38:00 crc kubenswrapper[4839]: E0227 19:38:00.969813 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:38:01 crc kubenswrapper[4839]: I0227 19:38:01.133276 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" event={"ID":"499a12e9-ae64-4d99-93ed-e6af775581c0","Type":"ContainerStarted","Data":"ac883ad2b4650d4ed67d251db2ca2bc803ab94aad7d17136ba7efb16b6b67ddd"} Feb 27 19:38:01 crc kubenswrapper[4839]: E0227 19:38:01.813924 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:38:01 crc kubenswrapper[4839]: E0227 19:38:01.814053 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:38:01 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:38:01 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j52c5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537018-dbqzs_openshift-infra(499a12e9-ae64-4d99-93ed-e6af775581c0): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:38:01 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:38:01 crc kubenswrapper[4839]: E0227 19:38:01.815233 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" podUID="499a12e9-ae64-4d99-93ed-e6af775581c0" Feb 27 19:38:01 crc kubenswrapper[4839]: E0227 19:38:01.966929 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:38:02 crc kubenswrapper[4839]: E0227 19:38:02.141094 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" podUID="499a12e9-ae64-4d99-93ed-e6af775581c0" Feb 27 19:38:02 crc kubenswrapper[4839]: E0227 19:38:02.967168 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:38:02 crc kubenswrapper[4839]: E0227 19:38:02.967581 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:38:05 crc kubenswrapper[4839]: E0227 19:38:05.966369 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jn87v" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.028313 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dk4jt" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.544127 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.544750 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.549056 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.549211 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.562697 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.646081 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93e64295-7c9f-43cb-9a2f-2a893e81e149-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.646149 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93e64295-7c9f-43cb-9a2f-2a893e81e149-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.747285 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93e64295-7c9f-43cb-9a2f-2a893e81e149-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.747373 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93e64295-7c9f-43cb-9a2f-2a893e81e149-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.747449 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93e64295-7c9f-43cb-9a2f-2a893e81e149-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.772219 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93e64295-7c9f-43cb-9a2f-2a893e81e149-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:06 crc kubenswrapper[4839]: I0227 19:38:06.879015 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:07 crc kubenswrapper[4839]: I0227 19:38:07.373823 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 19:38:07 crc kubenswrapper[4839]: W0227 19:38:07.379395 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod93e64295_7c9f_43cb_9a2f_2a893e81e149.slice/crio-dbc5981b133ed2e37031eb89a33101b33e1446d9e8048a66a115471c8d731a60 WatchSource:0}: Error finding container dbc5981b133ed2e37031eb89a33101b33e1446d9e8048a66a115471c8d731a60: Status 404 returned error can't find the container with id dbc5981b133ed2e37031eb89a33101b33e1446d9e8048a66a115471c8d731a60 Feb 27 19:38:07 crc kubenswrapper[4839]: E0227 19:38:07.968242 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-m6qmh" podUID="f11b8271-1de0-432d-8d78-2379862726a1" Feb 27 19:38:08 crc kubenswrapper[4839]: I0227 19:38:08.172896 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93e64295-7c9f-43cb-9a2f-2a893e81e149","Type":"ContainerStarted","Data":"c247155d1c9c5ad83ba1b39c03dda908582b210d16f89cf322b896db26fc687e"} Feb 27 19:38:08 crc kubenswrapper[4839]: I0227 19:38:08.173164 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93e64295-7c9f-43cb-9a2f-2a893e81e149","Type":"ContainerStarted","Data":"dbc5981b133ed2e37031eb89a33101b33e1446d9e8048a66a115471c8d731a60"} Feb 27 19:38:08 crc kubenswrapper[4839]: I0227 19:38:08.980315 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.980298942 podStartE2EDuration="2.980298942s" podCreationTimestamp="2026-02-27 19:38:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:08.188941632 +0000 UTC m=+249.833811367" watchObservedRunningTime="2026-02-27 19:38:08.980298942 +0000 UTC m=+250.625168677" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.178918 4839 generic.go:334] "Generic (PLEG): container finished" podID="93e64295-7c9f-43cb-9a2f-2a893e81e149" containerID="c247155d1c9c5ad83ba1b39c03dda908582b210d16f89cf322b896db26fc687e" exitCode=0 Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.178973 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93e64295-7c9f-43cb-9a2f-2a893e81e149","Type":"ContainerDied","Data":"c247155d1c9c5ad83ba1b39c03dda908582b210d16f89cf322b896db26fc687e"} Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.325017 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5967b47cf9-jvh6q"] Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.325283 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" podUID="eb84b0aa-f8fe-4074-aa11-13d147f52d52" containerName="controller-manager" containerID="cri-o://9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97" gracePeriod=30 Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.398513 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk"] Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.398731 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" podUID="038a5536-0e8e-478d-b743-be2c50c910fd" containerName="route-controller-manager" containerID="cri-o://b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310" gracePeriod=30 Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.826815 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.858210 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:38:09 crc kubenswrapper[4839]: E0227 19:38:09.951057 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:38:09 crc kubenswrapper[4839]: E0227 19:38:09.951212 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:38:09 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:38:09 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmg87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537016-rt648_openshift-infra(95e89fb2-3271-46bd-878b-6ff642b17214): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:38:09 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:38:09 crc kubenswrapper[4839]: E0227 19:38:09.952439 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.997225 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb84b0aa-f8fe-4074-aa11-13d147f52d52-serving-cert\") pod \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.997308 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4rdq\" (UniqueName: \"kubernetes.io/projected/eb84b0aa-f8fe-4074-aa11-13d147f52d52-kube-api-access-k4rdq\") pod \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.997374 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-client-ca\") pod \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.997422 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-client-ca\") pod \"038a5536-0e8e-478d-b743-be2c50c910fd\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.997507 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-config\") pod \"038a5536-0e8e-478d-b743-be2c50c910fd\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.998250 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-client-ca" (OuterVolumeSpecName: "client-ca") pod "eb84b0aa-f8fe-4074-aa11-13d147f52d52" (UID: "eb84b0aa-f8fe-4074-aa11-13d147f52d52"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.998271 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-client-ca" (OuterVolumeSpecName: "client-ca") pod "038a5536-0e8e-478d-b743-be2c50c910fd" (UID: "038a5536-0e8e-478d-b743-be2c50c910fd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.998281 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-config" (OuterVolumeSpecName: "config") pod "038a5536-0e8e-478d-b743-be2c50c910fd" (UID: "038a5536-0e8e-478d-b743-be2c50c910fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.998359 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-config\") pod \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.998402 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/038a5536-0e8e-478d-b743-be2c50c910fd-serving-cert\") pod \"038a5536-0e8e-478d-b743-be2c50c910fd\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.998898 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "eb84b0aa-f8fe-4074-aa11-13d147f52d52" (UID: "eb84b0aa-f8fe-4074-aa11-13d147f52d52"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999077 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-proxy-ca-bundles\") pod \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\" (UID: \"eb84b0aa-f8fe-4074-aa11-13d147f52d52\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999141 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr588\" (UniqueName: \"kubernetes.io/projected/038a5536-0e8e-478d-b743-be2c50c910fd-kube-api-access-zr588\") pod \"038a5536-0e8e-478d-b743-be2c50c910fd\" (UID: \"038a5536-0e8e-478d-b743-be2c50c910fd\") " Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999215 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-config" (OuterVolumeSpecName: "config") pod "eb84b0aa-f8fe-4074-aa11-13d147f52d52" (UID: "eb84b0aa-f8fe-4074-aa11-13d147f52d52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999459 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999476 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999489 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb84b0aa-f8fe-4074-aa11-13d147f52d52-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999500 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:09 crc kubenswrapper[4839]: I0227 19:38:09.999511 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038a5536-0e8e-478d-b743-be2c50c910fd-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.003046 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/038a5536-0e8e-478d-b743-be2c50c910fd-kube-api-access-zr588" (OuterVolumeSpecName: "kube-api-access-zr588") pod "038a5536-0e8e-478d-b743-be2c50c910fd" (UID: "038a5536-0e8e-478d-b743-be2c50c910fd"). InnerVolumeSpecName "kube-api-access-zr588". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.003100 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb84b0aa-f8fe-4074-aa11-13d147f52d52-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "eb84b0aa-f8fe-4074-aa11-13d147f52d52" (UID: "eb84b0aa-f8fe-4074-aa11-13d147f52d52"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.003173 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb84b0aa-f8fe-4074-aa11-13d147f52d52-kube-api-access-k4rdq" (OuterVolumeSpecName: "kube-api-access-k4rdq") pod "eb84b0aa-f8fe-4074-aa11-13d147f52d52" (UID: "eb84b0aa-f8fe-4074-aa11-13d147f52d52"). InnerVolumeSpecName "kube-api-access-k4rdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.003710 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/038a5536-0e8e-478d-b743-be2c50c910fd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "038a5536-0e8e-478d-b743-be2c50c910fd" (UID: "038a5536-0e8e-478d-b743-be2c50c910fd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.100395 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/038a5536-0e8e-478d-b743-be2c50c910fd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.100452 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr588\" (UniqueName: \"kubernetes.io/projected/038a5536-0e8e-478d-b743-be2c50c910fd-kube-api-access-zr588\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.100463 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb84b0aa-f8fe-4074-aa11-13d147f52d52-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.100472 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4rdq\" (UniqueName: \"kubernetes.io/projected/eb84b0aa-f8fe-4074-aa11-13d147f52d52-kube-api-access-k4rdq\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.185885 4839 generic.go:334] "Generic (PLEG): container finished" podID="eb84b0aa-f8fe-4074-aa11-13d147f52d52" containerID="9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97" exitCode=0 Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.185957 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" event={"ID":"eb84b0aa-f8fe-4074-aa11-13d147f52d52","Type":"ContainerDied","Data":"9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97"} Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.185960 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.185983 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967b47cf9-jvh6q" event={"ID":"eb84b0aa-f8fe-4074-aa11-13d147f52d52","Type":"ContainerDied","Data":"526af1c9e54c79ff3d15e1ef95bb057fbcc9b5b65ed627d6ab5062a22644cc98"} Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.186002 4839 scope.go:117] "RemoveContainer" containerID="9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.188827 4839 generic.go:334] "Generic (PLEG): container finished" podID="038a5536-0e8e-478d-b743-be2c50c910fd" containerID="b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310" exitCode=0 Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.188894 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.189521 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" event={"ID":"038a5536-0e8e-478d-b743-be2c50c910fd","Type":"ContainerDied","Data":"b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310"} Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.189555 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk" event={"ID":"038a5536-0e8e-478d-b743-be2c50c910fd","Type":"ContainerDied","Data":"0c0d368bb9cc30043d7832f8d16d2a7089f8d1c5581e99063279140d58b52def"} Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.205768 4839 scope.go:117] "RemoveContainer" containerID="9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97" Feb 27 19:38:10 crc kubenswrapper[4839]: E0227 19:38:10.206297 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97\": container with ID starting with 9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97 not found: ID does not exist" containerID="9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.206321 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97"} err="failed to get container status \"9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97\": rpc error: code = NotFound desc = could not find container \"9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97\": container with ID starting with 9ad8ecd78f0f7618323a4c8db0927a4bc4e89c7c03a747ee92e3c12faaaa3a97 not found: ID does not exist" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.206339 4839 scope.go:117] "RemoveContainer" containerID="b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.219375 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5967b47cf9-jvh6q"] Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.220207 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5967b47cf9-jvh6q"] Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.227988 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk"] Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.230500 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8449df6b88-c2wnk"] Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.231432 4839 scope.go:117] "RemoveContainer" containerID="b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310" Feb 27 19:38:10 crc kubenswrapper[4839]: E0227 19:38:10.232130 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310\": container with ID starting with b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310 not found: ID does not exist" containerID="b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.232158 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310"} err="failed to get container status \"b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310\": rpc error: code = NotFound desc = could not find container \"b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310\": container with ID starting with b5f568a816ded55a466b5c84480af9fdd832ad26347d07f3db25f71a20dea310 not found: ID does not exist" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.381597 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b448bbf4b-5jghc"] Feb 27 19:38:10 crc kubenswrapper[4839]: E0227 19:38:10.381993 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb84b0aa-f8fe-4074-aa11-13d147f52d52" containerName="controller-manager" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.382006 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb84b0aa-f8fe-4074-aa11-13d147f52d52" containerName="controller-manager" Feb 27 19:38:10 crc kubenswrapper[4839]: E0227 19:38:10.382028 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="038a5536-0e8e-478d-b743-be2c50c910fd" containerName="route-controller-manager" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.382034 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="038a5536-0e8e-478d-b743-be2c50c910fd" containerName="route-controller-manager" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.383112 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb84b0aa-f8fe-4074-aa11-13d147f52d52" containerName="controller-manager" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.383167 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="038a5536-0e8e-478d-b743-be2c50c910fd" containerName="route-controller-manager" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.383716 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.383724 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.387529 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.388863 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.389239 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.389752 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.389889 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.390431 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.391256 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b448bbf4b-5jghc"] Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.394869 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.406657 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.506517 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93e64295-7c9f-43cb-9a2f-2a893e81e149-kubelet-dir\") pod \"93e64295-7c9f-43cb-9a2f-2a893e81e149\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.506679 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93e64295-7c9f-43cb-9a2f-2a893e81e149-kube-api-access\") pod \"93e64295-7c9f-43cb-9a2f-2a893e81e149\" (UID: \"93e64295-7c9f-43cb-9a2f-2a893e81e149\") " Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.506858 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-config\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.506922 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m52wx\" (UniqueName: \"kubernetes.io/projected/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-kube-api-access-m52wx\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.506999 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-client-ca\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.507067 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e64295-7c9f-43cb-9a2f-2a893e81e149-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "93e64295-7c9f-43cb-9a2f-2a893e81e149" (UID: "93e64295-7c9f-43cb-9a2f-2a893e81e149"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.507137 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-serving-cert\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.507226 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-proxy-ca-bundles\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.507473 4839 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93e64295-7c9f-43cb-9a2f-2a893e81e149-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.510973 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e64295-7c9f-43cb-9a2f-2a893e81e149-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "93e64295-7c9f-43cb-9a2f-2a893e81e149" (UID: "93e64295-7c9f-43cb-9a2f-2a893e81e149"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.608632 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-client-ca\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.608721 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-serving-cert\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.608754 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-proxy-ca-bundles\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.608797 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-config\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.608831 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m52wx\" (UniqueName: \"kubernetes.io/projected/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-kube-api-access-m52wx\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.608886 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93e64295-7c9f-43cb-9a2f-2a893e81e149-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.610168 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-client-ca\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.610529 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-proxy-ca-bundles\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.610613 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-config\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.614376 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-serving-cert\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.625243 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m52wx\" (UniqueName: \"kubernetes.io/projected/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-kube-api-access-m52wx\") pod \"controller-manager-6b448bbf4b-5jghc\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.707427 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.971843 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="038a5536-0e8e-478d-b743-be2c50c910fd" path="/var/lib/kubelet/pods/038a5536-0e8e-478d-b743-be2c50c910fd/volumes" Feb 27 19:38:10 crc kubenswrapper[4839]: I0227 19:38:10.972454 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb84b0aa-f8fe-4074-aa11-13d147f52d52" path="/var/lib/kubelet/pods/eb84b0aa-f8fe-4074-aa11-13d147f52d52/volumes" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.082215 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b448bbf4b-5jghc"] Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.197059 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" event={"ID":"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe","Type":"ContainerStarted","Data":"87a9bf663943daa6b187ab589648fa9f12f6e7634f51134df4162b94e8e6bec6"} Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.198002 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.198740 4839 patch_prober.go:28] interesting pod/controller-manager-6b448bbf4b-5jghc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.198825 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" podUID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.199057 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93e64295-7c9f-43cb-9a2f-2a893e81e149","Type":"ContainerDied","Data":"dbc5981b133ed2e37031eb89a33101b33e1446d9e8048a66a115471c8d731a60"} Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.199084 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbc5981b133ed2e37031eb89a33101b33e1446d9e8048a66a115471c8d731a60" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.199110 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.214858 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" podStartSLOduration=2.214841895 podStartE2EDuration="2.214841895s" podCreationTimestamp="2026-02-27 19:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:11.212397272 +0000 UTC m=+252.857267017" watchObservedRunningTime="2026-02-27 19:38:11.214841895 +0000 UTC m=+252.859711630" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.381834 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9"] Feb 27 19:38:11 crc kubenswrapper[4839]: E0227 19:38:11.382101 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e64295-7c9f-43cb-9a2f-2a893e81e149" containerName="pruner" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.382117 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e64295-7c9f-43cb-9a2f-2a893e81e149" containerName="pruner" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.382234 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e64295-7c9f-43cb-9a2f-2a893e81e149" containerName="pruner" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.382910 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.385088 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.385173 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.385539 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.385574 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.385731 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.385831 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.395755 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9"] Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.520193 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-config\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.520243 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-client-ca\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.520311 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-serving-cert\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.520356 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kth9\" (UniqueName: \"kubernetes.io/projected/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-kube-api-access-6kth9\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.621197 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-serving-cert\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.621268 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kth9\" (UniqueName: \"kubernetes.io/projected/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-kube-api-access-6kth9\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.621312 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-config\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.621342 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-client-ca\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.622429 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-client-ca\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.622711 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-config\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.626223 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-serving-cert\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.645370 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kth9\" (UniqueName: \"kubernetes.io/projected/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-kube-api-access-6kth9\") pod \"route-controller-manager-66db5d5d98-5wrg9\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.699804 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:11 crc kubenswrapper[4839]: I0227 19:38:11.886816 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9"] Feb 27 19:38:11 crc kubenswrapper[4839]: W0227 19:38:11.890029 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc11d3aa4_ab81_44dd_a9e0_b28f1d1c7634.slice/crio-555b40ef1ba6c0601fbe3ba02d0f748ac53dc0617baf0926fdf2f4a0f5d8550b WatchSource:0}: Error finding container 555b40ef1ba6c0601fbe3ba02d0f748ac53dc0617baf0926fdf2f4a0f5d8550b: Status 404 returned error can't find the container with id 555b40ef1ba6c0601fbe3ba02d0f748ac53dc0617baf0926fdf2f4a0f5d8550b Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.204404 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" event={"ID":"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634","Type":"ContainerStarted","Data":"4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289"} Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.204633 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" event={"ID":"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634","Type":"ContainerStarted","Data":"555b40ef1ba6c0601fbe3ba02d0f748ac53dc0617baf0926fdf2f4a0f5d8550b"} Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.205075 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.206013 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" event={"ID":"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe","Type":"ContainerStarted","Data":"14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff"} Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.211427 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.222676 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" podStartSLOduration=3.222641418 podStartE2EDuration="3.222641418s" podCreationTimestamp="2026-02-27 19:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:12.22002067 +0000 UTC m=+253.864890405" watchObservedRunningTime="2026-02-27 19:38:12.222641418 +0000 UTC m=+253.867511153" Feb 27 19:38:12 crc kubenswrapper[4839]: I0227 19:38:12.465250 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.341324 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.342040 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.343591 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.349503 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.352542 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.443862 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-var-lock\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.444218 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.444256 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2056262a-25e6-4a59-badd-557fb91b48e4-kube-api-access\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.545273 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2056262a-25e6-4a59-badd-557fb91b48e4-kube-api-access\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.545367 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-var-lock\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.545395 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.545473 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.545531 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-var-lock\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.563421 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2056262a-25e6-4a59-badd-557fb91b48e4-kube-api-access\") pod \"installer-9-crc\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:13 crc kubenswrapper[4839]: I0227 19:38:13.662181 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:14 crc kubenswrapper[4839]: I0227 19:38:14.069227 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 19:38:14 crc kubenswrapper[4839]: W0227 19:38:14.077754 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2056262a_25e6_4a59_badd_557fb91b48e4.slice/crio-ba5631b8d72a2783234128e168964cfaf8f86d88d6facbd0ef0f090b7f46faab WatchSource:0}: Error finding container ba5631b8d72a2783234128e168964cfaf8f86d88d6facbd0ef0f090b7f46faab: Status 404 returned error can't find the container with id ba5631b8d72a2783234128e168964cfaf8f86d88d6facbd0ef0f090b7f46faab Feb 27 19:38:14 crc kubenswrapper[4839]: I0227 19:38:14.216488 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2056262a-25e6-4a59-badd-557fb91b48e4","Type":"ContainerStarted","Data":"ba5631b8d72a2783234128e168964cfaf8f86d88d6facbd0ef0f090b7f46faab"} Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.584720 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.584867 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n76bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ff888_openshift-marketplace(68a8a402-5d27-4632-8cf2-aaccccfff95d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.586087 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.607148 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.607477 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvsn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5p9bn_openshift-marketplace(8cc72c87-b6d6-4384-ac6e-fb12af841f66): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.608725 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.649550 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.649763 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8btrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-stz5c_openshift-marketplace(ff0a141c-3800-4c84-b1d6-17ec909d5d27): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:14 crc kubenswrapper[4839]: E0227 19:38:14.650938 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:38:15 crc kubenswrapper[4839]: I0227 19:38:15.224685 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2056262a-25e6-4a59-badd-557fb91b48e4","Type":"ContainerStarted","Data":"64b42202e3e2692ace5f748d96ffc68427edba4eb32a91400b9b59eeabf7a689"} Feb 27 19:38:15 crc kubenswrapper[4839]: I0227 19:38:15.242262 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.242225443 podStartE2EDuration="2.242225443s" podCreationTimestamp="2026-02-27 19:38:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:15.238214383 +0000 UTC m=+256.883084118" watchObservedRunningTime="2026-02-27 19:38:15.242225443 +0000 UTC m=+256.887095178" Feb 27 19:38:15 crc kubenswrapper[4839]: E0227 19:38:15.635166 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:38:15 crc kubenswrapper[4839]: E0227 19:38:15.635319 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w5zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7fddl_openshift-marketplace(67e667b9-304d-4052-8f0f-8d46d920aa7c): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:15 crc kubenswrapper[4839]: E0227 19:38:15.636584 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:38:15 crc kubenswrapper[4839]: E0227 19:38:15.706795 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:38:15 crc kubenswrapper[4839]: E0227 19:38:15.706972 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zm9fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgzcw_openshift-marketplace(0a4d0578-88dd-4754-81fd-ae28a709b698): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:15 crc kubenswrapper[4839]: E0227 19:38:15.708147 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:38:18 crc kubenswrapper[4839]: E0227 19:38:18.655981 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:38:18 crc kubenswrapper[4839]: E0227 19:38:18.656454 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5wrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d2p75_openshift-marketplace(b471c70a-15fe-4b5f-aa6f-354848925ace): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:18 crc kubenswrapper[4839]: E0227 19:38:18.657583 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:38:19 crc kubenswrapper[4839]: I0227 19:38:19.889093 4839 csr.go:261] certificate signing request csr-g8j85 is approved, waiting to be issued Feb 27 19:38:19 crc kubenswrapper[4839]: I0227 19:38:19.895256 4839 csr.go:257] certificate signing request csr-g8j85 is issued Feb 27 19:38:20 crc kubenswrapper[4839]: I0227 19:38:20.252818 4839 generic.go:334] "Generic (PLEG): container finished" podID="499a12e9-ae64-4d99-93ed-e6af775581c0" containerID="a7dbef3da3ac32c0108174826e5f63ebb0f520792390c542863b864ef66fea9c" exitCode=0 Feb 27 19:38:20 crc kubenswrapper[4839]: I0227 19:38:20.252857 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" event={"ID":"499a12e9-ae64-4d99-93ed-e6af775581c0","Type":"ContainerDied","Data":"a7dbef3da3ac32c0108174826e5f63ebb0f520792390c542863b864ef66fea9c"} Feb 27 19:38:20 crc kubenswrapper[4839]: I0227 19:38:20.896872 4839 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-13 16:35:27.877719924 +0000 UTC Feb 27 19:38:20 crc kubenswrapper[4839]: I0227 19:38:20.897207 4839 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7676h57m6.9805163s for next certificate rotation Feb 27 19:38:20 crc kubenswrapper[4839]: E0227 19:38:20.969000 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:38:21 crc kubenswrapper[4839]: I0227 19:38:21.586957 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:21 crc kubenswrapper[4839]: I0227 19:38:21.743147 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j52c5\" (UniqueName: \"kubernetes.io/projected/499a12e9-ae64-4d99-93ed-e6af775581c0-kube-api-access-j52c5\") pod \"499a12e9-ae64-4d99-93ed-e6af775581c0\" (UID: \"499a12e9-ae64-4d99-93ed-e6af775581c0\") " Feb 27 19:38:21 crc kubenswrapper[4839]: I0227 19:38:21.750024 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/499a12e9-ae64-4d99-93ed-e6af775581c0-kube-api-access-j52c5" (OuterVolumeSpecName: "kube-api-access-j52c5") pod "499a12e9-ae64-4d99-93ed-e6af775581c0" (UID: "499a12e9-ae64-4d99-93ed-e6af775581c0"). InnerVolumeSpecName "kube-api-access-j52c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:21 crc kubenswrapper[4839]: I0227 19:38:21.845085 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j52c5\" (UniqueName: \"kubernetes.io/projected/499a12e9-ae64-4d99-93ed-e6af775581c0-kube-api-access-j52c5\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:21 crc kubenswrapper[4839]: I0227 19:38:21.897647 4839 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-31 17:28:53.750579544 +0000 UTC Feb 27 19:38:21 crc kubenswrapper[4839]: I0227 19:38:21.897688 4839 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7365h50m31.852894267s for next certificate rotation Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.265057 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" event={"ID":"499a12e9-ae64-4d99-93ed-e6af775581c0","Type":"ContainerDied","Data":"ac883ad2b4650d4ed67d251db2ca2bc803ab94aad7d17136ba7efb16b6b67ddd"} Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.265105 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac883ad2b4650d4ed67d251db2ca2bc803ab94aad7d17136ba7efb16b6b67ddd" Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.265298 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537018-dbqzs" Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.682240 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.682318 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.682401 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.683114 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b"} pod="openshift-machine-config-operator/machine-config-daemon-thb8n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 19:38:22 crc kubenswrapper[4839]: I0227 19:38:22.683198 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" containerID="cri-o://f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b" gracePeriod=600 Feb 27 19:38:23 crc kubenswrapper[4839]: I0227 19:38:23.272158 4839 generic.go:334] "Generic (PLEG): container finished" podID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerID="f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b" exitCode=0 Feb 27 19:38:23 crc kubenswrapper[4839]: I0227 19:38:23.272261 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerDied","Data":"f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b"} Feb 27 19:38:26 crc kubenswrapper[4839]: E0227 19:38:26.380069 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.293005 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b448bbf4b-5jghc"] Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.293578 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" podUID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" containerName="controller-manager" containerID="cri-o://14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff" gracePeriod=30 Feb 27 19:38:29 crc kubenswrapper[4839]: E0227 19:38:29.295506 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:38:29 crc kubenswrapper[4839]: E0227 19:38:29.295572 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.315506 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9"] Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.315714 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" podUID="c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" containerName="route-controller-manager" containerID="cri-o://4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289" gracePeriod=30 Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.836456 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.849909 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949388 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-serving-cert\") pod \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949436 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-serving-cert\") pod \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949462 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-proxy-ca-bundles\") pod \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949496 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-client-ca\") pod \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949512 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-config\") pod \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949532 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m52wx\" (UniqueName: \"kubernetes.io/projected/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-kube-api-access-m52wx\") pod \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949573 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kth9\" (UniqueName: \"kubernetes.io/projected/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-kube-api-access-6kth9\") pod \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949593 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-config\") pod \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\" (UID: \"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.949624 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-client-ca\") pod \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\" (UID: \"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe\") " Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.950539 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-client-ca" (OuterVolumeSpecName: "client-ca") pod "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" (UID: "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.950619 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-config" (OuterVolumeSpecName: "config") pod "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" (UID: "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.950833 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" (UID: "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.950965 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-client-ca" (OuterVolumeSpecName: "client-ca") pod "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" (UID: "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.951119 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-config" (OuterVolumeSpecName: "config") pod "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" (UID: "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.954732 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-kube-api-access-6kth9" (OuterVolumeSpecName: "kube-api-access-6kth9") pod "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" (UID: "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634"). InnerVolumeSpecName "kube-api-access-6kth9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.955030 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" (UID: "c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.955155 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" (UID: "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: I0227 19:38:29.955190 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-kube-api-access-m52wx" (OuterVolumeSpecName: "kube-api-access-m52wx") pod "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" (UID: "9bf75073-30ba-4ff5-b235-3f6bd44ce4fe"). InnerVolumeSpecName "kube-api-access-m52wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:29 crc kubenswrapper[4839]: E0227 19:38:29.967020 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:38:29 crc kubenswrapper[4839]: E0227 19:38:29.967355 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051691 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051734 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051748 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051759 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051773 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051784 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051795 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m52wx\" (UniqueName: \"kubernetes.io/projected/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe-kube-api-access-m52wx\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051806 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kth9\" (UniqueName: \"kubernetes.io/projected/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-kube-api-access-6kth9\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.051816 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.318561 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerStarted","Data":"5d8d72fecee7144a5e8d81649356d93107c8fd3a59b133c077223d29f039f43b"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.322773 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"d66dbb8b18e6a5a083cdeabb414a44ff334ee6690e03ac1dc87884b44ab7d3f2"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.324309 4839 generic.go:334] "Generic (PLEG): container finished" podID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" containerID="14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff" exitCode=0 Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.324364 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" event={"ID":"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe","Type":"ContainerDied","Data":"14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.324393 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" event={"ID":"9bf75073-30ba-4ff5-b235-3f6bd44ce4fe","Type":"ContainerDied","Data":"87a9bf663943daa6b187ab589648fa9f12f6e7634f51134df4162b94e8e6bec6"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.324403 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b448bbf4b-5jghc" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.324421 4839 scope.go:117] "RemoveContainer" containerID="14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.326752 4839 generic.go:334] "Generic (PLEG): container finished" podID="c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" containerID="4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289" exitCode=0 Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.326852 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" event={"ID":"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634","Type":"ContainerDied","Data":"4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.326881 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" event={"ID":"c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634","Type":"ContainerDied","Data":"555b40ef1ba6c0601fbe3ba02d0f748ac53dc0617baf0926fdf2f4a0f5d8550b"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.326921 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.329780 4839 generic.go:334] "Generic (PLEG): container finished" podID="f11b8271-1de0-432d-8d78-2379862726a1" containerID="34b26e6b526bc8556137f5a0bcc2a8c24bc7556d1dc4bf20508e5d39ed9924d3" exitCode=0 Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.329811 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6qmh" event={"ID":"f11b8271-1de0-432d-8d78-2379862726a1","Type":"ContainerDied","Data":"34b26e6b526bc8556137f5a0bcc2a8c24bc7556d1dc4bf20508e5d39ed9924d3"} Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.404827 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-757f4b9549-lbh7f"] Feb 27 19:38:30 crc kubenswrapper[4839]: E0227 19:38:30.405194 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" containerName="route-controller-manager" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405213 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" containerName="route-controller-manager" Feb 27 19:38:30 crc kubenswrapper[4839]: E0227 19:38:30.405226 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499a12e9-ae64-4d99-93ed-e6af775581c0" containerName="oc" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405234 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="499a12e9-ae64-4d99-93ed-e6af775581c0" containerName="oc" Feb 27 19:38:30 crc kubenswrapper[4839]: E0227 19:38:30.405248 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" containerName="controller-manager" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405273 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" containerName="controller-manager" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405395 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" containerName="route-controller-manager" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405411 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="499a12e9-ae64-4d99-93ed-e6af775581c0" containerName="oc" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405422 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" containerName="controller-manager" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.405766 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.406336 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.406355 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.408014 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.411494 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.411745 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.411860 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.412540 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.412722 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-757f4b9549-lbh7f"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.411632 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.413403 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.413467 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.413488 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.413804 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.414009 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.414396 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.414517 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.417015 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.424614 4839 scope.go:117] "RemoveContainer" containerID="14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff" Feb 27 19:38:30 crc kubenswrapper[4839]: E0227 19:38:30.425040 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff\": container with ID starting with 14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff not found: ID does not exist" containerID="14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.425068 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff"} err="failed to get container status \"14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff\": rpc error: code = NotFound desc = could not find container \"14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff\": container with ID starting with 14450b02c9563cd924a2f21397602c36d3ced4ce9180129c05bc2580798355ff not found: ID does not exist" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.425089 4839 scope.go:117] "RemoveContainer" containerID="4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.459921 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b448bbf4b-5jghc"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467563 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k72kv\" (UniqueName: \"kubernetes.io/projected/b8bb9ba0-6677-4603-a15e-514a700e5bc0-kube-api-access-k72kv\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467686 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-config\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467733 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-client-ca\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467767 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5a3793-707f-471b-a5f8-63d8f1b46e28-serving-cert\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467813 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-config\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467847 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8bb9ba0-6677-4603-a15e-514a700e5bc0-serving-cert\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467883 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5lbg\" (UniqueName: \"kubernetes.io/projected/6e5a3793-707f-471b-a5f8-63d8f1b46e28-kube-api-access-q5lbg\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467926 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-client-ca\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.467950 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-proxy-ca-bundles\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.469355 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b448bbf4b-5jghc"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.470288 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.473202 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66db5d5d98-5wrg9"] Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.473503 4839 scope.go:117] "RemoveContainer" containerID="4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289" Feb 27 19:38:30 crc kubenswrapper[4839]: E0227 19:38:30.473942 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289\": container with ID starting with 4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289 not found: ID does not exist" containerID="4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.473984 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289"} err="failed to get container status \"4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289\": rpc error: code = NotFound desc = could not find container \"4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289\": container with ID starting with 4441633d920c9ac9be9ea5c53eb131c846e25e6aec4809abc82a5b5727c76289 not found: ID does not exist" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569605 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k72kv\" (UniqueName: \"kubernetes.io/projected/b8bb9ba0-6677-4603-a15e-514a700e5bc0-kube-api-access-k72kv\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569709 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-config\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569743 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-client-ca\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569773 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5a3793-707f-471b-a5f8-63d8f1b46e28-serving-cert\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569873 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-config\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569910 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8bb9ba0-6677-4603-a15e-514a700e5bc0-serving-cert\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569940 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5lbg\" (UniqueName: \"kubernetes.io/projected/6e5a3793-707f-471b-a5f8-63d8f1b46e28-kube-api-access-q5lbg\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.569976 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-client-ca\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.570000 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-proxy-ca-bundles\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.571561 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-client-ca\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.571621 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-proxy-ca-bundles\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.571929 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-config\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.571945 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-client-ca\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.572049 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-config\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.575386 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5a3793-707f-471b-a5f8-63d8f1b46e28-serving-cert\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.580213 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8bb9ba0-6677-4603-a15e-514a700e5bc0-serving-cert\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.586159 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5lbg\" (UniqueName: \"kubernetes.io/projected/6e5a3793-707f-471b-a5f8-63d8f1b46e28-kube-api-access-q5lbg\") pod \"controller-manager-757f4b9549-lbh7f\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.586797 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k72kv\" (UniqueName: \"kubernetes.io/projected/b8bb9ba0-6677-4603-a15e-514a700e5bc0-kube-api-access-k72kv\") pod \"route-controller-manager-84dfd68899-xqh9m\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.774290 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.781472 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:30 crc kubenswrapper[4839]: E0227 19:38:30.966757 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.980194 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bf75073-30ba-4ff5-b235-3f6bd44ce4fe" path="/var/lib/kubelet/pods/9bf75073-30ba-4ff5-b235-3f6bd44ce4fe/volumes" Feb 27 19:38:30 crc kubenswrapper[4839]: I0227 19:38:30.980975 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634" path="/var/lib/kubelet/pods/c11d3aa4-ab81-44dd-a9e0-b28f1d1c7634/volumes" Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.157564 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m"] Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.203250 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-757f4b9549-lbh7f"] Feb 27 19:38:31 crc kubenswrapper[4839]: W0227 19:38:31.211005 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e5a3793_707f_471b_a5f8_63d8f1b46e28.slice/crio-4e4f9588149e3ca6369c2f574e859a3703d28760b81a78698302b2efde687307 WatchSource:0}: Error finding container 4e4f9588149e3ca6369c2f574e859a3703d28760b81a78698302b2efde687307: Status 404 returned error can't find the container with id 4e4f9588149e3ca6369c2f574e859a3703d28760b81a78698302b2efde687307 Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.337907 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6qmh" event={"ID":"f11b8271-1de0-432d-8d78-2379862726a1","Type":"ContainerStarted","Data":"13d27d812bf2e04ba7894fd3d40f834e035ea6cc1d6d6bd4adb237611ec7d7c4"} Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.339873 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" event={"ID":"6e5a3793-707f-471b-a5f8-63d8f1b46e28","Type":"ContainerStarted","Data":"e963271d282aff56a6641bc618b01f445bd6b5c46436c5dde4608c146825cbb6"} Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.340151 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" event={"ID":"6e5a3793-707f-471b-a5f8-63d8f1b46e28","Type":"ContainerStarted","Data":"4e4f9588149e3ca6369c2f574e859a3703d28760b81a78698302b2efde687307"} Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.340177 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.341831 4839 patch_prober.go:28] interesting pod/controller-manager-757f4b9549-lbh7f container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.341879 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" podUID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.341905 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerDied","Data":"5d8d72fecee7144a5e8d81649356d93107c8fd3a59b133c077223d29f039f43b"} Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.341884 4839 generic.go:334] "Generic (PLEG): container finished" podID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerID="5d8d72fecee7144a5e8d81649356d93107c8fd3a59b133c077223d29f039f43b" exitCode=0 Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.351461 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" event={"ID":"b8bb9ba0-6677-4603-a15e-514a700e5bc0","Type":"ContainerStarted","Data":"66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6"} Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.351513 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" event={"ID":"b8bb9ba0-6677-4603-a15e-514a700e5bc0","Type":"ContainerStarted","Data":"fc4450b636eef119456ca1e8eb1a55011033538dfe2c254e058801960219055f"} Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.358681 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m6qmh" podStartSLOduration=2.373711197 podStartE2EDuration="57.358648923s" podCreationTimestamp="2026-02-27 19:37:34 +0000 UTC" firstStartedPulling="2026-02-27 19:37:35.957017582 +0000 UTC m=+217.601887337" lastFinishedPulling="2026-02-27 19:38:30.941955328 +0000 UTC m=+272.586825063" observedRunningTime="2026-02-27 19:38:31.357455552 +0000 UTC m=+273.002325287" watchObservedRunningTime="2026-02-27 19:38:31.358648923 +0000 UTC m=+273.003518658" Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.392733 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" podStartSLOduration=2.3920101 podStartE2EDuration="2.3920101s" podCreationTimestamp="2026-02-27 19:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:31.389959106 +0000 UTC m=+273.034828851" watchObservedRunningTime="2026-02-27 19:38:31.3920101 +0000 UTC m=+273.036879835" Feb 27 19:38:31 crc kubenswrapper[4839]: I0227 19:38:31.407952 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" podStartSLOduration=2.407933883 podStartE2EDuration="2.407933883s" podCreationTimestamp="2026-02-27 19:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:31.406529346 +0000 UTC m=+273.051399091" watchObservedRunningTime="2026-02-27 19:38:31.407933883 +0000 UTC m=+273.052803618" Feb 27 19:38:32 crc kubenswrapper[4839]: I0227 19:38:32.360522 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerStarted","Data":"fc5bdb83191312b55b11b604f9a761620a6dfbd6bc67121090d3c8e5f96c3849"} Feb 27 19:38:32 crc kubenswrapper[4839]: I0227 19:38:32.360874 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:32 crc kubenswrapper[4839]: I0227 19:38:32.366467 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:32 crc kubenswrapper[4839]: I0227 19:38:32.368097 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:32 crc kubenswrapper[4839]: I0227 19:38:32.381456 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jn87v" podStartSLOduration=1.583622789 podStartE2EDuration="57.38144179s" podCreationTimestamp="2026-02-27 19:37:35 +0000 UTC" firstStartedPulling="2026-02-27 19:37:35.960586958 +0000 UTC m=+217.605456693" lastFinishedPulling="2026-02-27 19:38:31.758405959 +0000 UTC m=+273.403275694" observedRunningTime="2026-02-27 19:38:32.38030639 +0000 UTC m=+274.025176145" watchObservedRunningTime="2026-02-27 19:38:32.38144179 +0000 UTC m=+274.026311525" Feb 27 19:38:34 crc kubenswrapper[4839]: I0227 19:38:34.222395 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v6dfv"] Feb 27 19:38:34 crc kubenswrapper[4839]: E0227 19:38:34.967900 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:38:35 crc kubenswrapper[4839]: I0227 19:38:35.189518 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:38:35 crc kubenswrapper[4839]: I0227 19:38:35.189575 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:38:35 crc kubenswrapper[4839]: I0227 19:38:35.587059 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:38:35 crc kubenswrapper[4839]: I0227 19:38:35.587201 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:38:36 crc kubenswrapper[4839]: I0227 19:38:36.302357 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m6qmh" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="registry-server" probeResult="failure" output=< Feb 27 19:38:36 crc kubenswrapper[4839]: timeout: failed to connect service ":50051" within 1s Feb 27 19:38:36 crc kubenswrapper[4839]: > Feb 27 19:38:36 crc kubenswrapper[4839]: I0227 19:38:36.626444 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jn87v" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="registry-server" probeResult="failure" output=< Feb 27 19:38:36 crc kubenswrapper[4839]: timeout: failed to connect service ":50051" within 1s Feb 27 19:38:36 crc kubenswrapper[4839]: > Feb 27 19:38:39 crc kubenswrapper[4839]: E0227 19:38:39.969210 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" Feb 27 19:38:40 crc kubenswrapper[4839]: E0227 19:38:40.967116 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:38:43 crc kubenswrapper[4839]: E0227 19:38:43.967601 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:38:43 crc kubenswrapper[4839]: E0227 19:38:43.967625 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:38:43 crc kubenswrapper[4839]: E0227 19:38:43.967775 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:38:45 crc kubenswrapper[4839]: I0227 19:38:45.283851 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:38:45 crc kubenswrapper[4839]: I0227 19:38:45.321446 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:38:45 crc kubenswrapper[4839]: I0227 19:38:45.625616 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:38:45 crc kubenswrapper[4839]: I0227 19:38:45.668460 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:38:45 crc kubenswrapper[4839]: E0227 19:38:45.966935 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:38:46 crc kubenswrapper[4839]: I0227 19:38:46.410840 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jn87v"] Feb 27 19:38:47 crc kubenswrapper[4839]: I0227 19:38:47.432425 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jn87v" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="registry-server" containerID="cri-o://fc5bdb83191312b55b11b604f9a761620a6dfbd6bc67121090d3c8e5f96c3849" gracePeriod=2 Feb 27 19:38:48 crc kubenswrapper[4839]: E0227 19:38:48.968826 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537016-rt648" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.299152 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-757f4b9549-lbh7f"] Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.299387 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" podUID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" containerName="controller-manager" containerID="cri-o://e963271d282aff56a6641bc618b01f445bd6b5c46436c5dde4608c146825cbb6" gracePeriod=30 Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.400520 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m"] Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.400753 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" podUID="b8bb9ba0-6677-4603-a15e-514a700e5bc0" containerName="route-controller-manager" containerID="cri-o://66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6" gracePeriod=30 Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.443808 4839 generic.go:334] "Generic (PLEG): container finished" podID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" containerID="e963271d282aff56a6641bc618b01f445bd6b5c46436c5dde4608c146825cbb6" exitCode=0 Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.443871 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" event={"ID":"6e5a3793-707f-471b-a5f8-63d8f1b46e28","Type":"ContainerDied","Data":"e963271d282aff56a6641bc618b01f445bd6b5c46436c5dde4608c146825cbb6"} Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.445600 4839 generic.go:334] "Generic (PLEG): container finished" podID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerID="fc5bdb83191312b55b11b604f9a761620a6dfbd6bc67121090d3c8e5f96c3849" exitCode=0 Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.445629 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerDied","Data":"fc5bdb83191312b55b11b604f9a761620a6dfbd6bc67121090d3c8e5f96c3849"} Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.961250 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:49 crc kubenswrapper[4839]: I0227 19:38:49.983559 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.002797 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8bb9ba0-6677-4603-a15e-514a700e5bc0-serving-cert\") pod \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.003252 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-proxy-ca-bundles\") pod \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.004299 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-client-ca" (OuterVolumeSpecName: "client-ca") pod "6e5a3793-707f-471b-a5f8-63d8f1b46e28" (UID: "6e5a3793-707f-471b-a5f8-63d8f1b46e28"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.004712 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6e5a3793-707f-471b-a5f8-63d8f1b46e28" (UID: "6e5a3793-707f-471b-a5f8-63d8f1b46e28"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.003284 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-client-ca\") pod \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.005645 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5a3793-707f-471b-a5f8-63d8f1b46e28-serving-cert\") pod \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.006069 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-client-ca\") pod \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.007412 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-client-ca" (OuterVolumeSpecName: "client-ca") pod "b8bb9ba0-6677-4603-a15e-514a700e5bc0" (UID: "b8bb9ba0-6677-4603-a15e-514a700e5bc0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.007565 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-config\") pod \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.007591 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k72kv\" (UniqueName: \"kubernetes.io/projected/b8bb9ba0-6677-4603-a15e-514a700e5bc0-kube-api-access-k72kv\") pod \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\" (UID: \"b8bb9ba0-6677-4603-a15e-514a700e5bc0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.007610 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8bb9ba0-6677-4603-a15e-514a700e5bc0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b8bb9ba0-6677-4603-a15e-514a700e5bc0" (UID: "b8bb9ba0-6677-4603-a15e-514a700e5bc0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008166 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-config\") pod \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008194 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5lbg\" (UniqueName: \"kubernetes.io/projected/6e5a3793-707f-471b-a5f8-63d8f1b46e28-kube-api-access-q5lbg\") pod \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\" (UID: \"6e5a3793-707f-471b-a5f8-63d8f1b46e28\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008766 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-config" (OuterVolumeSpecName: "config") pod "6e5a3793-707f-471b-a5f8-63d8f1b46e28" (UID: "6e5a3793-707f-471b-a5f8-63d8f1b46e28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008806 4839 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008822 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8bb9ba0-6677-4603-a15e-514a700e5bc0-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008831 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008839 4839 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.008191 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-config" (OuterVolumeSpecName: "config") pod "b8bb9ba0-6677-4603-a15e-514a700e5bc0" (UID: "b8bb9ba0-6677-4603-a15e-514a700e5bc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.015046 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5a3793-707f-471b-a5f8-63d8f1b46e28-kube-api-access-q5lbg" (OuterVolumeSpecName: "kube-api-access-q5lbg") pod "6e5a3793-707f-471b-a5f8-63d8f1b46e28" (UID: "6e5a3793-707f-471b-a5f8-63d8f1b46e28"). InnerVolumeSpecName "kube-api-access-q5lbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.015240 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8bb9ba0-6677-4603-a15e-514a700e5bc0-kube-api-access-k72kv" (OuterVolumeSpecName: "kube-api-access-k72kv") pod "b8bb9ba0-6677-4603-a15e-514a700e5bc0" (UID: "b8bb9ba0-6677-4603-a15e-514a700e5bc0"). InnerVolumeSpecName "kube-api-access-k72kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.015328 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5a3793-707f-471b-a5f8-63d8f1b46e28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6e5a3793-707f-471b-a5f8-63d8f1b46e28" (UID: "6e5a3793-707f-471b-a5f8-63d8f1b46e28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.018041 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110114 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-utilities\") pod \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110165 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-catalog-content\") pod \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110235 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9svsj\" (UniqueName: \"kubernetes.io/projected/6321a495-e9eb-4ca0-8adf-af955a1a87c0-kube-api-access-9svsj\") pod \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\" (UID: \"6321a495-e9eb-4ca0-8adf-af955a1a87c0\") " Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110440 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8bb9ba0-6677-4603-a15e-514a700e5bc0-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110452 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k72kv\" (UniqueName: \"kubernetes.io/projected/b8bb9ba0-6677-4603-a15e-514a700e5bc0-kube-api-access-k72kv\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110463 4839 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a3793-707f-471b-a5f8-63d8f1b46e28-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110472 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5lbg\" (UniqueName: \"kubernetes.io/projected/6e5a3793-707f-471b-a5f8-63d8f1b46e28-kube-api-access-q5lbg\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.110480 4839 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5a3793-707f-471b-a5f8-63d8f1b46e28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.111149 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-utilities" (OuterVolumeSpecName: "utilities") pod "6321a495-e9eb-4ca0-8adf-af955a1a87c0" (UID: "6321a495-e9eb-4ca0-8adf-af955a1a87c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.113080 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6321a495-e9eb-4ca0-8adf-af955a1a87c0-kube-api-access-9svsj" (OuterVolumeSpecName: "kube-api-access-9svsj") pod "6321a495-e9eb-4ca0-8adf-af955a1a87c0" (UID: "6321a495-e9eb-4ca0-8adf-af955a1a87c0"). InnerVolumeSpecName "kube-api-access-9svsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.211437 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.211469 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9svsj\" (UniqueName: \"kubernetes.io/projected/6321a495-e9eb-4ca0-8adf-af955a1a87c0-kube-api-access-9svsj\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.240754 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6321a495-e9eb-4ca0-8adf-af955a1a87c0" (UID: "6321a495-e9eb-4ca0-8adf-af955a1a87c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.312145 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6321a495-e9eb-4ca0-8adf-af955a1a87c0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.415955 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-fcfbc464b-s7p25"] Feb 27 19:38:50 crc kubenswrapper[4839]: E0227 19:38:50.416208 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" containerName="controller-manager" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416228 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" containerName="controller-manager" Feb 27 19:38:50 crc kubenswrapper[4839]: E0227 19:38:50.416239 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="extract-utilities" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416248 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="extract-utilities" Feb 27 19:38:50 crc kubenswrapper[4839]: E0227 19:38:50.416265 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="extract-content" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416274 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="extract-content" Feb 27 19:38:50 crc kubenswrapper[4839]: E0227 19:38:50.416293 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bb9ba0-6677-4603-a15e-514a700e5bc0" containerName="route-controller-manager" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416301 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bb9ba0-6677-4603-a15e-514a700e5bc0" containerName="route-controller-manager" Feb 27 19:38:50 crc kubenswrapper[4839]: E0227 19:38:50.416316 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="registry-server" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416324 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="registry-server" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416441 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" containerName="controller-manager" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416456 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8bb9ba0-6677-4603-a15e-514a700e5bc0" containerName="route-controller-manager" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.416482 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" containerName="registry-server" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.417075 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.432358 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-fcfbc464b-s7p25"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.453273 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.453359 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-757f4b9549-lbh7f" event={"ID":"6e5a3793-707f-471b-a5f8-63d8f1b46e28","Type":"ContainerDied","Data":"4e4f9588149e3ca6369c2f574e859a3703d28760b81a78698302b2efde687307"} Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.453449 4839 scope.go:117] "RemoveContainer" containerID="e963271d282aff56a6641bc618b01f445bd6b5c46436c5dde4608c146825cbb6" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.457603 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn87v" event={"ID":"6321a495-e9eb-4ca0-8adf-af955a1a87c0","Type":"ContainerDied","Data":"50e30c9a5fd7cf82eb6f29728385ee1c045c45bc11efb9f14e239ab2547c14ec"} Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.457754 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn87v" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.459574 4839 generic.go:334] "Generic (PLEG): container finished" podID="b8bb9ba0-6677-4603-a15e-514a700e5bc0" containerID="66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6" exitCode=0 Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.459619 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" event={"ID":"b8bb9ba0-6677-4603-a15e-514a700e5bc0","Type":"ContainerDied","Data":"66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6"} Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.459650 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" event={"ID":"b8bb9ba0-6677-4603-a15e-514a700e5bc0","Type":"ContainerDied","Data":"fc4450b636eef119456ca1e8eb1a55011033538dfe2c254e058801960219055f"} Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.459704 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.475402 4839 scope.go:117] "RemoveContainer" containerID="fc5bdb83191312b55b11b604f9a761620a6dfbd6bc67121090d3c8e5f96c3849" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.482600 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-757f4b9549-lbh7f"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.487262 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-757f4b9549-lbh7f"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.496710 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jn87v"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.499933 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jn87v"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.507654 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.507813 4839 scope.go:117] "RemoveContainer" containerID="5d8d72fecee7144a5e8d81649356d93107c8fd3a59b133c077223d29f039f43b" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.511809 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dfd68899-xqh9m"] Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.515135 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-client-ca\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.515194 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-proxy-ca-bundles\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.515216 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gccs\" (UniqueName: \"kubernetes.io/projected/85249e2b-89c5-4119-baa0-95021fdfb453-kube-api-access-8gccs\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.515257 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-config\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.515282 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85249e2b-89c5-4119-baa0-95021fdfb453-serving-cert\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.523893 4839 scope.go:117] "RemoveContainer" containerID="f5e10603dc3d7f8858cdaf045d7be6d1b373e231e3f6d783d7341eb4d6266e33" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.538013 4839 scope.go:117] "RemoveContainer" containerID="66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.555503 4839 scope.go:117] "RemoveContainer" containerID="66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6" Feb 27 19:38:50 crc kubenswrapper[4839]: E0227 19:38:50.556028 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6\": container with ID starting with 66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6 not found: ID does not exist" containerID="66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.556061 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6"} err="failed to get container status \"66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6\": rpc error: code = NotFound desc = could not find container \"66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6\": container with ID starting with 66cff0369bf5934a628cb97715d2ef262b7a2b42544f728626424334a23723e6 not found: ID does not exist" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.616781 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gccs\" (UniqueName: \"kubernetes.io/projected/85249e2b-89c5-4119-baa0-95021fdfb453-kube-api-access-8gccs\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.616854 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-config\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.616976 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85249e2b-89c5-4119-baa0-95021fdfb453-serving-cert\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.617002 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-client-ca\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.617040 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-proxy-ca-bundles\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.618185 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-client-ca\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.618201 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-config\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.618309 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85249e2b-89c5-4119-baa0-95021fdfb453-proxy-ca-bundles\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.629324 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85249e2b-89c5-4119-baa0-95021fdfb453-serving-cert\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.636261 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gccs\" (UniqueName: \"kubernetes.io/projected/85249e2b-89c5-4119-baa0-95021fdfb453-kube-api-access-8gccs\") pod \"controller-manager-fcfbc464b-s7p25\" (UID: \"85249e2b-89c5-4119-baa0-95021fdfb453\") " pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.753619 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.970896 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6321a495-e9eb-4ca0-8adf-af955a1a87c0" path="/var/lib/kubelet/pods/6321a495-e9eb-4ca0-8adf-af955a1a87c0/volumes" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.971508 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e5a3793-707f-471b-a5f8-63d8f1b46e28" path="/var/lib/kubelet/pods/6e5a3793-707f-471b-a5f8-63d8f1b46e28/volumes" Feb 27 19:38:50 crc kubenswrapper[4839]: I0227 19:38:50.971999 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8bb9ba0-6677-4603-a15e-514a700e5bc0" path="/var/lib/kubelet/pods/b8bb9ba0-6677-4603-a15e-514a700e5bc0/volumes" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.168377 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-fcfbc464b-s7p25"] Feb 27 19:38:51 crc kubenswrapper[4839]: W0227 19:38:51.169626 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85249e2b_89c5_4119_baa0_95021fdfb453.slice/crio-417af8272d3e6955ecf030529d780a909e67b7cf1ee05b95ff3b391fe3588da0 WatchSource:0}: Error finding container 417af8272d3e6955ecf030529d780a909e67b7cf1ee05b95ff3b391fe3588da0: Status 404 returned error can't find the container with id 417af8272d3e6955ecf030529d780a909e67b7cf1ee05b95ff3b391fe3588da0 Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.412644 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh"] Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.413431 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.416036 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.416145 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.416498 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.416548 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.416622 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.417485 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.424605 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh"] Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.467581 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" event={"ID":"85249e2b-89c5-4119-baa0-95021fdfb453","Type":"ContainerStarted","Data":"5a9fb44ccd31eb5048146a60d4e1d59b38c5667055f2552311dcdec57ec4eacf"} Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.467628 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" event={"ID":"85249e2b-89c5-4119-baa0-95021fdfb453","Type":"ContainerStarted","Data":"417af8272d3e6955ecf030529d780a909e67b7cf1ee05b95ff3b391fe3588da0"} Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.467958 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.480079 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.491068 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-fcfbc464b-s7p25" podStartSLOduration=2.491048512 podStartE2EDuration="2.491048512s" podCreationTimestamp="2026-02-27 19:38:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:38:51.489805109 +0000 UTC m=+293.134674874" watchObservedRunningTime="2026-02-27 19:38:51.491048512 +0000 UTC m=+293.135918247" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.529173 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f403baf0-1e43-479e-be29-06503d46ceee-serving-cert\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.529226 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlq5x\" (UniqueName: \"kubernetes.io/projected/f403baf0-1e43-479e-be29-06503d46ceee-kube-api-access-qlq5x\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.529305 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f403baf0-1e43-479e-be29-06503d46ceee-client-ca\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.529337 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f403baf0-1e43-479e-be29-06503d46ceee-config\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.630551 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f403baf0-1e43-479e-be29-06503d46ceee-client-ca\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.630824 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f403baf0-1e43-479e-be29-06503d46ceee-config\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.630962 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f403baf0-1e43-479e-be29-06503d46ceee-serving-cert\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.631076 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlq5x\" (UniqueName: \"kubernetes.io/projected/f403baf0-1e43-479e-be29-06503d46ceee-kube-api-access-qlq5x\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.631749 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f403baf0-1e43-479e-be29-06503d46ceee-config\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.632206 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f403baf0-1e43-479e-be29-06503d46ceee-client-ca\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.635894 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f403baf0-1e43-479e-be29-06503d46ceee-serving-cert\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.653219 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlq5x\" (UniqueName: \"kubernetes.io/projected/f403baf0-1e43-479e-be29-06503d46ceee-kube-api-access-qlq5x\") pod \"route-controller-manager-5499f549c8-5l5qh\" (UID: \"f403baf0-1e43-479e-be29-06503d46ceee\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.747600 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:51 crc kubenswrapper[4839]: I0227 19:38:51.957775 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh"] Feb 27 19:38:51 crc kubenswrapper[4839]: W0227 19:38:51.962475 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf403baf0_1e43_479e_be29_06503d46ceee.slice/crio-fcfbc937f4a9fce80e536671bf87f2f0e6d06fa301f0e2cc26f49bdef57917b1 WatchSource:0}: Error finding container fcfbc937f4a9fce80e536671bf87f2f0e6d06fa301f0e2cc26f49bdef57917b1: Status 404 returned error can't find the container with id fcfbc937f4a9fce80e536671bf87f2f0e6d06fa301f0e2cc26f49bdef57917b1 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.009602 4839 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.010295 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.010592 4839 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.011301 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430" gracePeriod=15 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.011379 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad" gracePeriod=15 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.011366 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e" gracePeriod=15 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.011323 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18" gracePeriod=15 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.011379 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd" gracePeriod=15 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012310 4839 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012624 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012649 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012691 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012702 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012713 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012722 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012734 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012742 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012753 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012760 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012771 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012786 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012797 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012802 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012809 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012815 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.012827 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012833 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012939 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012948 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012954 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012962 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012970 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012978 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012985 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.012995 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.013078 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.013085 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.013181 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036369 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036435 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036479 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036540 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036573 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036598 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036673 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.036712 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.080771 4839 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.89:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: E0227 19:38:52.095954 4839 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/events\": dial tcp 38.102.83.89:6443: connect: connection refused" event="&Event{ObjectMeta:{route-controller-manager-5499f549c8-5l5qh.189831bae1fd0048 openshift-route-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-route-controller-manager,Name:route-controller-manager-5499f549c8-5l5qh,UID:f403baf0-1e43-479e-be29-06503d46ceee,APIVersion:v1,ResourceVersion:30033,FieldPath:spec.containers{route-controller-manager},},Reason:Created,Message:Created container route-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:38:52.095422536 +0000 UTC m=+293.740292271,LastTimestamp:2026-02-27 19:38:52.095422536 +0000 UTC m=+293.740292271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137513 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137589 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137617 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137602 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137646 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137696 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137712 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137711 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137777 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137749 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137764 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137850 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137860 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.137829 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.138058 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.138138 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.381723 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:52 crc kubenswrapper[4839]: W0227 19:38:52.400709 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-a88241a94cc04273d4b293b502f31a975ed0c18945371afbc50be9814572ae56 WatchSource:0}: Error finding container a88241a94cc04273d4b293b502f31a975ed0c18945371afbc50be9814572ae56: Status 404 returned error can't find the container with id a88241a94cc04273d4b293b502f31a975ed0c18945371afbc50be9814572ae56 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.478352 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.479560 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.480286 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e" exitCode=0 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.480318 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18" exitCode=0 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.480330 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd" exitCode=0 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.480340 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad" exitCode=2 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.480384 4839 scope.go:117] "RemoveContainer" containerID="c6a3763e726b560d56fdec610432cfac222c833b283d35b17e178c08bed4df4f" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.482844 4839 generic.go:334] "Generic (PLEG): container finished" podID="2056262a-25e6-4a59-badd-557fb91b48e4" containerID="64b42202e3e2692ace5f748d96ffc68427edba4eb32a91400b9b59eeabf7a689" exitCode=0 Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.482900 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2056262a-25e6-4a59-badd-557fb91b48e4","Type":"ContainerDied","Data":"64b42202e3e2692ace5f748d96ffc68427edba4eb32a91400b9b59eeabf7a689"} Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.483444 4839 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.483707 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.484776 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a88241a94cc04273d4b293b502f31a975ed0c18945371afbc50be9814572ae56"} Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.486617 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" event={"ID":"f403baf0-1e43-479e-be29-06503d46ceee","Type":"ContainerStarted","Data":"ed61fa889898eee76b18bd27fb94f887f19ba47fdbb86bb36797519c9f7ed67d"} Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.486657 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" event={"ID":"f403baf0-1e43-479e-be29-06503d46ceee","Type":"ContainerStarted","Data":"fcfbc937f4a9fce80e536671bf87f2f0e6d06fa301f0e2cc26f49bdef57917b1"} Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.487341 4839 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.488091 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:52 crc kubenswrapper[4839]: I0227 19:38:52.488492 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.452961 4839 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.453835 4839 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.454201 4839 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.454541 4839 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.454865 4839 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.454899 4839 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.455288 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="200ms" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.495800 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.498106 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1"} Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.498652 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.498939 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.499403 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.499437 4839 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.89:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.656809 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="400ms" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.794439 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.794954 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.795341 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.863416 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2056262a-25e6-4a59-badd-557fb91b48e4-kube-api-access\") pod \"2056262a-25e6-4a59-badd-557fb91b48e4\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.863508 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-kubelet-dir\") pod \"2056262a-25e6-4a59-badd-557fb91b48e4\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.863559 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-var-lock\") pod \"2056262a-25e6-4a59-badd-557fb91b48e4\" (UID: \"2056262a-25e6-4a59-badd-557fb91b48e4\") " Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.863624 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2056262a-25e6-4a59-badd-557fb91b48e4" (UID: "2056262a-25e6-4a59-badd-557fb91b48e4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.863815 4839 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.863862 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-var-lock" (OuterVolumeSpecName: "var-lock") pod "2056262a-25e6-4a59-badd-557fb91b48e4" (UID: "2056262a-25e6-4a59-badd-557fb91b48e4"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.871226 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2056262a-25e6-4a59-badd-557fb91b48e4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2056262a-25e6-4a59-badd-557fb91b48e4" (UID: "2056262a-25e6-4a59-badd-557fb91b48e4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.965118 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2056262a-25e6-4a59-badd-557fb91b48e4-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:53 crc kubenswrapper[4839]: I0227 19:38:53.965279 4839 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2056262a-25e6-4a59-badd-557fb91b48e4-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:53 crc kubenswrapper[4839]: E0227 19:38:53.966567 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.057760 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="800ms" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.379109 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.380733 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.381604 4839 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.381845 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.382031 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470219 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470345 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470339 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470370 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470400 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470530 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470780 4839 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470793 4839 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.470802 4839 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.498905 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.499189 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.509578 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.512153 4839 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430" exitCode=0 Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.512274 4839 scope.go:117] "RemoveContainer" containerID="2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.512345 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.514068 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2056262a-25e6-4a59-badd-557fb91b48e4","Type":"ContainerDied","Data":"ba5631b8d72a2783234128e168964cfaf8f86d88d6facbd0ef0f090b7f46faab"} Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.514137 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba5631b8d72a2783234128e168964cfaf8f86d88d6facbd0ef0f090b7f46faab" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.514177 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.515518 4839 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.89:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.533933 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.534424 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.535018 4839 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.535509 4839 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.535761 4839 scope.go:117] "RemoveContainer" containerID="b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.535914 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.536977 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.552715 4839 scope.go:117] "RemoveContainer" containerID="c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.568048 4839 scope.go:117] "RemoveContainer" containerID="8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.582765 4839 scope.go:117] "RemoveContainer" containerID="0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.610461 4839 scope.go:117] "RemoveContainer" containerID="9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.633706 4839 scope.go:117] "RemoveContainer" containerID="2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.634874 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\": container with ID starting with 2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e not found: ID does not exist" containerID="2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.634921 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e"} err="failed to get container status \"2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\": rpc error: code = NotFound desc = could not find container \"2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e\": container with ID starting with 2cf6b584d9aafe7d976eb7ee8cb137f97cda48721afd5b7a69d464b6993f098e not found: ID does not exist" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.634954 4839 scope.go:117] "RemoveContainer" containerID="b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.635327 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\": container with ID starting with b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18 not found: ID does not exist" containerID="b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.635461 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18"} err="failed to get container status \"b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\": rpc error: code = NotFound desc = could not find container \"b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18\": container with ID starting with b5af13859f9865fc451b52a06bd83a0d31fd0a6744f0ae39bcedb1fb2a281b18 not found: ID does not exist" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.635571 4839 scope.go:117] "RemoveContainer" containerID="c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.636314 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\": container with ID starting with c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd not found: ID does not exist" containerID="c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.636344 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd"} err="failed to get container status \"c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\": rpc error: code = NotFound desc = could not find container \"c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd\": container with ID starting with c5c9098ed1c671f84ed7533a0dd2829959ba54553751f05ed2b9426566e743cd not found: ID does not exist" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.636364 4839 scope.go:117] "RemoveContainer" containerID="8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.636726 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\": container with ID starting with 8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad not found: ID does not exist" containerID="8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.636778 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad"} err="failed to get container status \"8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\": rpc error: code = NotFound desc = could not find container \"8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad\": container with ID starting with 8e76d4f640febe7ebdeda9fc467596143c24ea9df8c9dc01a43830009e168bad not found: ID does not exist" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.636813 4839 scope.go:117] "RemoveContainer" containerID="0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.637126 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\": container with ID starting with 0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430 not found: ID does not exist" containerID="0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.637156 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430"} err="failed to get container status \"0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\": rpc error: code = NotFound desc = could not find container \"0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430\": container with ID starting with 0d2882d71bdd046f0d933d27e35265a25998a92156e801bf980ec2f8854b1430 not found: ID does not exist" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.637175 4839 scope.go:117] "RemoveContainer" containerID="9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.637409 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\": container with ID starting with 9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2 not found: ID does not exist" containerID="9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.637443 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2"} err="failed to get container status \"9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\": rpc error: code = NotFound desc = could not find container \"9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2\": container with ID starting with 9a59954595b44e3e4df762ca4b1f5099a4743cdace212b818d4bbf6267db5cc2 not found: ID does not exist" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.858865 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="1.6s" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.968299 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" Feb 27 19:38:54 crc kubenswrapper[4839]: E0227 19:38:54.968328 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:38:54 crc kubenswrapper[4839]: I0227 19:38:54.979505 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 27 19:38:55 crc kubenswrapper[4839]: I0227 19:38:55.515171 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:38:55 crc kubenswrapper[4839]: I0227 19:38:55.515446 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:38:55 crc kubenswrapper[4839]: E0227 19:38:55.973554 4839 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.89:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" volumeName="registry-storage" Feb 27 19:38:56 crc kubenswrapper[4839]: E0227 19:38:56.459585 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="3.2s" Feb 27 19:38:56 crc kubenswrapper[4839]: E0227 19:38:56.967594 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:38:58 crc kubenswrapper[4839]: E0227 19:38:58.791181 4839 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/events\": dial tcp 38.102.83.89:6443: connect: connection refused" event="&Event{ObjectMeta:{route-controller-manager-5499f549c8-5l5qh.189831bae1fd0048 openshift-route-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-route-controller-manager,Name:route-controller-manager-5499f549c8-5l5qh,UID:f403baf0-1e43-479e-be29-06503d46ceee,APIVersion:v1,ResourceVersion:30033,FieldPath:spec.containers{route-controller-manager},},Reason:Created,Message:Created container route-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 19:38:52.095422536 +0000 UTC m=+293.740292271,LastTimestamp:2026-02-27 19:38:52.095422536 +0000 UTC m=+293.740292271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 19:38:58 crc kubenswrapper[4839]: I0227 19:38:58.967439 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:58 crc kubenswrapper[4839]: I0227 19:38:58.967677 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.266409 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" containerName="oauth-openshift" containerID="cri-o://d4f00d348616a22c368cfc002710a199d73219051ee3f2cb32e4f337e77c4005" gracePeriod=15 Feb 27 19:38:59 crc kubenswrapper[4839]: E0227 19:38:59.417690 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:38:59 crc kubenswrapper[4839]: E0227 19:38:59.417889 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n76bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ff888_openshift-marketplace(68a8a402-5d27-4632-8cf2-aaccccfff95d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:38:59 crc kubenswrapper[4839]: E0227 19:38:59.419696 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.557845 4839 generic.go:334] "Generic (PLEG): container finished" podID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" containerID="d4f00d348616a22c368cfc002710a199d73219051ee3f2cb32e4f337e77c4005" exitCode=0 Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.557950 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" event={"ID":"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c","Type":"ContainerDied","Data":"d4f00d348616a22c368cfc002710a199d73219051ee3f2cb32e4f337e77c4005"} Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.560704 4839 generic.go:334] "Generic (PLEG): container finished" podID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerID="d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2" exitCode=0 Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.560761 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5p9bn" event={"ID":"8cc72c87-b6d6-4384-ac6e-fb12af841f66","Type":"ContainerDied","Data":"d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2"} Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.561781 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.562214 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.566781 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: E0227 19:38:59.660405 4839 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" interval="6.4s" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.750118 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.750938 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.751494 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.752101 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.752651 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.852748 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvzgj\" (UniqueName: \"kubernetes.io/projected/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-kube-api-access-hvzgj\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.852820 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-service-ca\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.852871 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-login\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.852895 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-serving-cert\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853730 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-session\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853765 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-trusted-ca-bundle\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853792 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-router-certs\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853819 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-policies\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853814 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853851 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-cliconfig\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853881 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-ocp-branding-template\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853930 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-error\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.853965 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-provider-selection\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.854031 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-dir\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.854053 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-idp-0-file-data\") pod \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\" (UID: \"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c\") " Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.854261 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.854361 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.854553 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.854842 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.855013 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.858644 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.858709 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-kube-api-access-hvzgj" (OuterVolumeSpecName: "kube-api-access-hvzgj") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "kube-api-access-hvzgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.859259 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.863078 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.866079 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.866963 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.867177 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.867372 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.867508 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" (UID: "93e80011-4bd5-4ce0-ac3f-38ca6c67a63c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955732 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955791 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955813 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955834 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955854 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955874 4839 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955893 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955913 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955932 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955951 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955971 4839 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.955992 4839 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 19:38:59 crc kubenswrapper[4839]: I0227 19:38:59.956014 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvzgj\" (UniqueName: \"kubernetes.io/projected/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c-kube-api-access-hvzgj\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.569688 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5p9bn" event={"ID":"8cc72c87-b6d6-4384-ac6e-fb12af841f66","Type":"ContainerStarted","Data":"6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3"} Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.571173 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.571843 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.572247 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" event={"ID":"93e80011-4bd5-4ce0-ac3f-38ca6c67a63c","Type":"ContainerDied","Data":"6df1b11c412c6bedd75d7a6c733af77084876f4f766f30e711db6c64659e8cd9"} Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.572279 4839 scope.go:117] "RemoveContainer" containerID="d4f00d348616a22c368cfc002710a199d73219051ee3f2cb32e4f337e77c4005" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.572338 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.572238 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.572741 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.573048 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.573285 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.573746 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.574119 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.597942 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.598507 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.599296 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:00 crc kubenswrapper[4839]: I0227 19:39:00.600133 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.411923 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.412261 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.452928 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.453501 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.453900 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.454203 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.454459 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.590709 4839 generic.go:334] "Generic (PLEG): container finished" podID="95e89fb2-3271-46bd-878b-6ff642b17214" containerID="4091d93ec5f4af7327331baae1e03bbc9c9124f98cf2ba8eb41266edf46e5f29" exitCode=0 Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.590813 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537016-rt648" event={"ID":"95e89fb2-3271-46bd-878b-6ff642b17214","Type":"ContainerDied","Data":"4091d93ec5f4af7327331baae1e03bbc9c9124f98cf2ba8eb41266edf46e5f29"} Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.591550 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.591905 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.592184 4839 status_manager.go:851] "Failed to get status for pod" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" pod="openshift-infra/auto-csr-approver-29537016-rt648" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29537016-rt648\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.592394 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.592630 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.748840 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:39:02 crc kubenswrapper[4839]: I0227 19:39:02.748904 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.964622 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.965759 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.966082 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.966612 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.967364 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.967877 4839 status_manager.go:851] "Failed to get status for pod" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" pod="openshift-infra/auto-csr-approver-29537016-rt648" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29537016-rt648\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.970110 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.970798 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.971434 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.971988 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.972644 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.974311 4839 status_manager.go:851] "Failed to get status for pod" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" pod="openshift-infra/auto-csr-approver-29537016-rt648" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29537016-rt648\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.978597 4839 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.978625 4839 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:03 crc kubenswrapper[4839]: E0227 19:39:03.978974 4839 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:03 crc kubenswrapper[4839]: I0227 19:39:03.979386 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.005585 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmg87\" (UniqueName: \"kubernetes.io/projected/95e89fb2-3271-46bd-878b-6ff642b17214-kube-api-access-kmg87\") pod \"95e89fb2-3271-46bd-878b-6ff642b17214\" (UID: \"95e89fb2-3271-46bd-878b-6ff642b17214\") " Feb 27 19:39:04 crc kubenswrapper[4839]: W0227 19:39:04.007047 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-ecb2db58ca4beadfa78fa7f14eec48d3ad2e6d8466c9d26f7abe3c065352de65 WatchSource:0}: Error finding container ecb2db58ca4beadfa78fa7f14eec48d3ad2e6d8466c9d26f7abe3c065352de65: Status 404 returned error can't find the container with id ecb2db58ca4beadfa78fa7f14eec48d3ad2e6d8466c9d26f7abe3c065352de65 Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.010955 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e89fb2-3271-46bd-878b-6ff642b17214-kube-api-access-kmg87" (OuterVolumeSpecName: "kube-api-access-kmg87") pod "95e89fb2-3271-46bd-878b-6ff642b17214" (UID: "95e89fb2-3271-46bd-878b-6ff642b17214"). InnerVolumeSpecName "kube-api-access-kmg87". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.106788 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmg87\" (UniqueName: \"kubernetes.io/projected/95e89fb2-3271-46bd-878b-6ff642b17214-kube-api-access-kmg87\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.609087 4839 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="614fb91390764bdbcdf9c7269447ab917c5fb46df16a00c752c4341bf958d1c9" exitCode=0 Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.609971 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"614fb91390764bdbcdf9c7269447ab917c5fb46df16a00c752c4341bf958d1c9"} Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.610150 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ecb2db58ca4beadfa78fa7f14eec48d3ad2e6d8466c9d26f7abe3c065352de65"} Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.610651 4839 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.610885 4839 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.611812 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.611952 4839 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.612845 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.613465 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.613875 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.614207 4839 status_manager.go:851] "Failed to get status for pod" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" pod="openshift-infra/auto-csr-approver-29537016-rt648" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29537016-rt648\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.616031 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537016-rt648" event={"ID":"95e89fb2-3271-46bd-878b-6ff642b17214","Type":"ContainerDied","Data":"58c003c295f585d416eb192851977ffecd334166767f3ebf3f304da6448aa3f7"} Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.616082 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58c003c295f585d416eb192851977ffecd334166767f3ebf3f304da6448aa3f7" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.616104 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537016-rt648" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.635993 4839 status_manager.go:851] "Failed to get status for pod" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" pod="openshift-authentication/oauth-openshift-558db77b4-v6dfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v6dfv\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.637107 4839 status_manager.go:851] "Failed to get status for pod" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.643975 4839 status_manager.go:851] "Failed to get status for pod" podUID="f403baf0-1e43-479e-be29-06503d46ceee" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5499f549c8-5l5qh\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.645010 4839 status_manager.go:851] "Failed to get status for pod" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" pod="openshift-marketplace/certified-operators-5p9bn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5p9bn\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: I0227 19:39:04.646373 4839 status_manager.go:851] "Failed to get status for pod" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" pod="openshift-infra/auto-csr-approver-29537016-rt648" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29537016-rt648\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.901633 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:39:04Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:39:04Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:39:04Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T19:39:04Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:73193fb63f256beb1da0cc4095be1a794952cc273142e451a0475e479e7cc56d\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1709186004},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:220e38c2edfe1a5e22b9f251de4149a79b780a8ecba94fe02f1508f5331ed512\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1265396955},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.902102 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.902503 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.902770 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.902936 4839 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.89:6443: connect: connection refused" Feb 27 19:39:04 crc kubenswrapper[4839]: E0227 19:39:04.902955 4839 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 19:39:05 crc kubenswrapper[4839]: I0227 19:39:05.643975 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ab869efcf30d4580eaf7af8e1369f41110bf7520bb833f0e1cce27430ec5bbc1"} Feb 27 19:39:05 crc kubenswrapper[4839]: I0227 19:39:05.644251 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"855860113b672db7eefde9faf96940027fbcd833c1f5a20d51f5d8c588a89157"} Feb 27 19:39:05 crc kubenswrapper[4839]: I0227 19:39:05.644265 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d38fcd52524b6d95544443fa17d3d4d344df858b4b478a1c760b4a7209236072"} Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.651259 4839 generic.go:334] "Generic (PLEG): container finished" podID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerID="2f9e7c5935162e0ff84fc261b0127931500b22f6d1e95825491f070d9d629c4b" exitCode=0 Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.651325 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-stz5c" event={"ID":"ff0a141c-3800-4c84-b1d6-17ec909d5d27","Type":"ContainerDied","Data":"2f9e7c5935162e0ff84fc261b0127931500b22f6d1e95825491f070d9d629c4b"} Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.655824 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e121b1c020728fa7a8e2251e569a65259379804fa044735465ff0df1616e57d8"} Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.655920 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b97837d9c9f8496bb9c471397f8a16149c1369e30e0aa8fb4391caf50ee51f0e"} Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.656043 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.656161 4839 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:06 crc kubenswrapper[4839]: I0227 19:39:06.656197 4839 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.505010 4839 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.505345 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.663161 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.664132 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.664257 4839 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3" exitCode=1 Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.664356 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3"} Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.664960 4839 scope.go:117] "RemoveContainer" containerID="0614491d1bd477c567246ba3adb9055f58d5d0f0546e4ded684efb0b781756e3" Feb 27 19:39:07 crc kubenswrapper[4839]: I0227 19:39:07.667052 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-stz5c" event={"ID":"ff0a141c-3800-4c84-b1d6-17ec909d5d27","Type":"ContainerStarted","Data":"c062e48e1dcfbe0524b8bf04be3045ea174c4a5b81b1200ae146399c4db1bc29"} Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.241186 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:39:08 crc kubenswrapper[4839]: E0227 19:39:08.528769 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:39:08 crc kubenswrapper[4839]: E0227 19:39:08.528924 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zm9fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgzcw_openshift-marketplace(0a4d0578-88dd-4754-81fd-ae28a709b698): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:39:08 crc kubenswrapper[4839]: E0227 19:39:08.530177 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.676530 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.677279 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.677348 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5695da9f1fe48d1101616bb4911c4f289dab8f03d9674fcce43b4bd5c149a9c5"} Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.979917 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.980248 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:08 crc kubenswrapper[4839]: I0227 19:39:08.987826 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:09 crc kubenswrapper[4839]: I0227 19:39:09.548156 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:39:11 crc kubenswrapper[4839]: E0227 19:39:11.654433 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 19:39:11 crc kubenswrapper[4839]: E0227 19:39:11.654840 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5wrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d2p75_openshift-marketplace(b471c70a-15fe-4b5f-aa6f-354848925ace): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:39:11 crc kubenswrapper[4839]: E0227 19:39:11.656454 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:39:11 crc kubenswrapper[4839]: I0227 19:39:11.664652 4839 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:11 crc kubenswrapper[4839]: I0227 19:39:11.693458 4839 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:11 crc kubenswrapper[4839]: I0227 19:39:11.693491 4839 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:11 crc kubenswrapper[4839]: I0227 19:39:11.699980 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:11 crc kubenswrapper[4839]: I0227 19:39:11.785282 4839 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5af8a36f-76bb-4a90-89c7-48cbc0eec11c" Feb 27 19:39:11 crc kubenswrapper[4839]: E0227 19:39:11.966993 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.006256 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.006315 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.053281 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.450221 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.701832 4839 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.703472 4839 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2c04a518-2635-44d2-bff2-1332fd913a23" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.705060 4839 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5af8a36f-76bb-4a90-89c7-48cbc0eec11c" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.742147 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.748963 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:39:12 crc kubenswrapper[4839]: I0227 19:39:12.749017 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:39:14 crc kubenswrapper[4839]: I0227 19:39:14.714398 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerStarted","Data":"740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c"} Feb 27 19:39:15 crc kubenswrapper[4839]: I0227 19:39:15.724264 4839 generic.go:334] "Generic (PLEG): container finished" podID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerID="740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c" exitCode=0 Feb 27 19:39:15 crc kubenswrapper[4839]: I0227 19:39:15.724455 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerDied","Data":"740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c"} Feb 27 19:39:16 crc kubenswrapper[4839]: I0227 19:39:16.733145 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerStarted","Data":"f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1"} Feb 27 19:39:18 crc kubenswrapper[4839]: I0227 19:39:18.240742 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:39:18 crc kubenswrapper[4839]: I0227 19:39:18.241803 4839 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 27 19:39:18 crc kubenswrapper[4839]: I0227 19:39:18.241875 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 27 19:39:21 crc kubenswrapper[4839]: I0227 19:39:21.347103 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 19:39:21 crc kubenswrapper[4839]: E0227 19:39:21.968863 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.273075 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.301114 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.398863 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": read tcp 10.217.0.2:42484->10.217.0.68:8443: read: connection reset by peer" start-of-body= Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.399022 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": read tcp 10.217.0.2:42484->10.217.0.68:8443: read: connection reset by peer" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.517087 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.545454 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.771853 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-5499f549c8-5l5qh_f403baf0-1e43-479e-be29-06503d46ceee/route-controller-manager/0.log" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.772190 4839 generic.go:334] "Generic (PLEG): container finished" podID="f403baf0-1e43-479e-be29-06503d46ceee" containerID="ed61fa889898eee76b18bd27fb94f887f19ba47fdbb86bb36797519c9f7ed67d" exitCode=255 Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.772296 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" event={"ID":"f403baf0-1e43-479e-be29-06503d46ceee","Type":"ContainerDied","Data":"ed61fa889898eee76b18bd27fb94f887f19ba47fdbb86bb36797519c9f7ed67d"} Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.772848 4839 scope.go:117] "RemoveContainer" containerID="ed61fa889898eee76b18bd27fb94f887f19ba47fdbb86bb36797519c9f7ed67d" Feb 27 19:39:22 crc kubenswrapper[4839]: E0227 19:39:22.966457 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:39:22 crc kubenswrapper[4839]: I0227 19:39:22.968414 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 19:39:23 crc kubenswrapper[4839]: I0227 19:39:23.073315 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 19:39:23 crc kubenswrapper[4839]: I0227 19:39:23.475347 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 19:39:23 crc kubenswrapper[4839]: I0227 19:39:23.779764 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-5499f549c8-5l5qh_f403baf0-1e43-479e-be29-06503d46ceee/route-controller-manager/0.log" Feb 27 19:39:23 crc kubenswrapper[4839]: I0227 19:39:23.779836 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" event={"ID":"f403baf0-1e43-479e-be29-06503d46ceee","Type":"ContainerStarted","Data":"1f3fe29855ee30cc1d67446a40fe5b0782fdef99a54409448b9994a54580db3e"} Feb 27 19:39:23 crc kubenswrapper[4839]: I0227 19:39:23.780208 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.038970 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.077126 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.189899 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.189962 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.228849 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.300941 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.370204 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.480471 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.574749 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.768212 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.780707 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.780772 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.839786 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.937532 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 19:39:24 crc kubenswrapper[4839]: E0227 19:39:24.970285 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:39:24 crc kubenswrapper[4839]: I0227 19:39:24.977350 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.038437 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.057966 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.079585 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.158793 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.373913 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.464983 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.623407 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.642126 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.787798 4839 patch_prober.go:28] interesting pod/route-controller-manager-5499f549c8-5l5qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 19:39:25 crc kubenswrapper[4839]: I0227 19:39:25.787874 4839 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podUID="f403baf0-1e43-479e-be29-06503d46ceee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.067031 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.143873 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.250950 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.251098 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.429559 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.486856 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.614639 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.670160 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.788708 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.788814 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.802523 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.861856 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.897243 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.983423 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 19:39:26 crc kubenswrapper[4839]: I0227 19:39:26.987622 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.126872 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.171561 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.182605 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.222798 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.231252 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.274241 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.308162 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.313756 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.349614 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.560021 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.694229 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.717009 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.739379 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.750118 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.750593 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.795002 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.869980 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.909442 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 19:39:27 crc kubenswrapper[4839]: I0227 19:39:27.928950 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.017893 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.075111 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.087440 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.097338 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.124090 4839 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.246395 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.251388 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.256003 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.345006 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.358774 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.394313 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.438817 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.487607 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.531057 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.710946 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.840780 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.891142 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.986308 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 19:39:28 crc kubenswrapper[4839]: I0227 19:39:28.998023 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.157936 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.279337 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.391561 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.569197 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.579420 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.689797 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.750551 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.761128 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.779303 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 19:39:29 crc kubenswrapper[4839]: I0227 19:39:29.865653 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.030616 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.128232 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.140827 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.141647 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.168193 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.262895 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.307309 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.314935 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.346016 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.426606 4839 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.440330 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.460031 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.461280 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.533301 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.540326 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.722089 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.736159 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.771644 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.790423 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.796924 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.847044 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.856991 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.914972 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.975184 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 19:39:30 crc kubenswrapper[4839]: I0227 19:39:30.975468 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.008450 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.025350 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.071514 4839 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.083432 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.135654 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.272882 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.294827 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.296544 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.398382 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.405962 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.428068 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.477606 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.569878 4839 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.658209 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.683216 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.711174 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.744314 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.752425 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.776830 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.824851 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 19:39:31 crc kubenswrapper[4839]: I0227 19:39:31.950011 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.000831 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.199049 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.218172 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.335777 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.440360 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.505462 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.544083 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.568584 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.650093 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.681113 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.706784 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.725679 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.818145 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 19:39:32 crc kubenswrapper[4839]: I0227 19:39:32.889732 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 19:39:32 crc kubenswrapper[4839]: E0227 19:39:32.967630 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.074259 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.080406 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.133742 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.198138 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.216409 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.262017 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.392183 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.417408 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.441419 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.467469 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.525098 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.566848 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.579629 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.603248 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.746701 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.747056 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.810892 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.830744 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.884784 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.903635 4839 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.904778 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5p9bn" podStartSLOduration=35.750004114 podStartE2EDuration="2m1.90475574s" podCreationTimestamp="2026-02-27 19:37:32 +0000 UTC" firstStartedPulling="2026-02-27 19:37:33.867787066 +0000 UTC m=+215.512656801" lastFinishedPulling="2026-02-27 19:39:00.022538692 +0000 UTC m=+301.667408427" observedRunningTime="2026-02-27 19:39:11.682524799 +0000 UTC m=+313.327394554" watchObservedRunningTime="2026-02-27 19:39:33.90475574 +0000 UTC m=+335.549625515" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.904950 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7fddl" podStartSLOduration=19.666880537 podStartE2EDuration="2m0.904944031s" podCreationTimestamp="2026-02-27 19:37:33 +0000 UTC" firstStartedPulling="2026-02-27 19:37:34.913035112 +0000 UTC m=+216.557904847" lastFinishedPulling="2026-02-27 19:39:16.151098606 +0000 UTC m=+317.795968341" observedRunningTime="2026-02-27 19:39:16.746161504 +0000 UTC m=+318.391031239" watchObservedRunningTime="2026-02-27 19:39:33.904944031 +0000 UTC m=+335.549813806" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.907338 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-5l5qh" podStartSLOduration=44.907325724 podStartE2EDuration="44.907325724s" podCreationTimestamp="2026-02-27 19:38:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:39:11.771930626 +0000 UTC m=+313.416800361" watchObservedRunningTime="2026-02-27 19:39:33.907325724 +0000 UTC m=+335.552195499" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.908578 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-stz5c" podStartSLOduration=29.659483506 podStartE2EDuration="2m2.908567273s" podCreationTimestamp="2026-02-27 19:37:31 +0000 UTC" firstStartedPulling="2026-02-27 19:37:33.813987905 +0000 UTC m=+215.458857640" lastFinishedPulling="2026-02-27 19:39:07.063071642 +0000 UTC m=+308.707941407" observedRunningTime="2026-02-27 19:39:11.802686303 +0000 UTC m=+313.447556038" watchObservedRunningTime="2026-02-27 19:39:33.908567273 +0000 UTC m=+335.553437058" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.911560 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-v6dfv"] Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.911636 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.917911 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.933083 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.933059034 podStartE2EDuration="22.933059034s" podCreationTimestamp="2026-02-27 19:39:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:39:33.932510431 +0000 UTC m=+335.577380166" watchObservedRunningTime="2026-02-27 19:39:33.933059034 +0000 UTC m=+335.577928809" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.960835 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 19:39:33 crc kubenswrapper[4839]: E0227 19:39:33.967007 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:39:33 crc kubenswrapper[4839]: I0227 19:39:33.998254 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.011260 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.071924 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.082568 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.109847 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.218574 4839 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.218846 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1" gracePeriod=5 Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.253930 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.303883 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.315570 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.439853 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.490458 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.571953 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.657805 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.684375 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.724380 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.823462 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.874963 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.876125 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.978798 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 19:39:34 crc kubenswrapper[4839]: I0227 19:39:34.981368 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" path="/var/lib/kubelet/pods/93e80011-4bd5-4ce0-ac3f-38ca6c67a63c/volumes" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.021517 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.028628 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.043320 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.058480 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.075070 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.160320 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.230987 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.325413 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.376540 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.461468 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.656460 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.691540 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 19:39:35 crc kubenswrapper[4839]: I0227 19:39:35.906974 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.016520 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.102781 4839 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.104502 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.105930 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.122117 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.365350 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.385991 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.438012 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.448680 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.465160 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.741271 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.780417 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.803293 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 19:39:36 crc kubenswrapper[4839]: I0227 19:39:36.805230 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.196229 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.227467 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.263828 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.265331 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.277967 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.302196 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.356297 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.433909 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.705455 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.784371 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.870119 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 19:39:37 crc kubenswrapper[4839]: I0227 19:39:37.979121 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.168207 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.527173 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.574276 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.672164 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.718179 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.732382 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 19:39:38 crc kubenswrapper[4839]: I0227 19:39:38.860474 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 19:39:38 crc kubenswrapper[4839]: E0227 19:39:38.973612 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.010024 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.549537 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.596189 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.609512 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.798504 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.798618 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.875576 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.875928 4839 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1" exitCode=137 Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.876050 4839 scope.go:117] "RemoveContainer" containerID="fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.876045 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.892651 4839 scope.go:117] "RemoveContainer" containerID="fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1" Feb 27 19:39:39 crc kubenswrapper[4839]: E0227 19:39:39.893193 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1\": container with ID starting with fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1 not found: ID does not exist" containerID="fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.893232 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1"} err="failed to get container status \"fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1\": rpc error: code = NotFound desc = could not find container \"fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1\": container with ID starting with fef02d3c052eb3fababd7619d2eee050547cded129d63526ab75c3bd9ef4d7a1 not found: ID does not exist" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894044 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894117 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894208 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894183 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894362 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894382 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894446 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894462 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894620 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894908 4839 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894951 4839 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894964 4839 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.894975 4839 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.907376 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.907366 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:39:39 crc kubenswrapper[4839]: I0227 19:39:39.996840 4839 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.189979 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.444802 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9645b9d-w9zp7"] Feb 27 19:39:40 crc kubenswrapper[4839]: E0227 19:39:40.445244 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.445321 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 19:39:40 crc kubenswrapper[4839]: E0227 19:39:40.445385 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" containerName="installer" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.445443 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" containerName="installer" Feb 27 19:39:40 crc kubenswrapper[4839]: E0227 19:39:40.445499 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" containerName="oc" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.445554 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" containerName="oc" Feb 27 19:39:40 crc kubenswrapper[4839]: E0227 19:39:40.445785 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" containerName="oauth-openshift" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.445857 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" containerName="oauth-openshift" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.446032 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.446106 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="2056262a-25e6-4a59-badd-557fb91b48e4" containerName="installer" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.446168 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e80011-4bd5-4ce0-ac3f-38ca6c67a63c" containerName="oauth-openshift" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.446228 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" containerName="oc" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.446858 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.448908 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.451120 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.451256 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.451772 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.451845 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.452009 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.452059 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.452221 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.452373 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.452577 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.452639 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.453680 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.460850 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.462003 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9645b9d-w9zp7"] Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.465555 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.472180 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503303 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503353 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-router-certs\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503377 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503408 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503470 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8xks\" (UniqueName: \"kubernetes.io/projected/99311d9f-5bd3-4280-a720-26be8b87b47f-kube-api-access-k8xks\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503522 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503590 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503628 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-session\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503682 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-service-ca\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503724 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-audit-policies\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503764 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503800 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99311d9f-5bd3-4280-a720-26be8b87b47f-audit-dir\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503825 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-error\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.503854 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-login\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605258 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605466 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605512 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-session\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605532 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-service-ca\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605565 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-audit-policies\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605598 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605631 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99311d9f-5bd3-4280-a720-26be8b87b47f-audit-dir\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605659 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-error\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605727 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-login\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605752 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605783 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-router-certs\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605808 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605844 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.605873 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8xks\" (UniqueName: \"kubernetes.io/projected/99311d9f-5bd3-4280-a720-26be8b87b47f-kube-api-access-k8xks\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.606342 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.607418 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.607489 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99311d9f-5bd3-4280-a720-26be8b87b47f-audit-dir\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.607625 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-audit-policies\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.607908 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-service-ca\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.610342 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.610437 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-login\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.611298 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.611307 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.611837 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-error\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.613365 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-router-certs\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.619143 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.622317 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/99311d9f-5bd3-4280-a720-26be8b87b47f-v4-0-config-system-session\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.632981 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8xks\" (UniqueName: \"kubernetes.io/projected/99311d9f-5bd3-4280-a720-26be8b87b47f-kube-api-access-k8xks\") pod \"oauth-openshift-9645b9d-w9zp7\" (UID: \"99311d9f-5bd3-4280-a720-26be8b87b47f\") " pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.765558 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.920607 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 19:39:40 crc kubenswrapper[4839]: I0227 19:39:40.972006 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 27 19:39:41 crc kubenswrapper[4839]: I0227 19:39:41.172259 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9645b9d-w9zp7"] Feb 27 19:39:41 crc kubenswrapper[4839]: W0227 19:39:41.175853 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99311d9f_5bd3_4280_a720_26be8b87b47f.slice/crio-94d11e220810737144e8281d689a2d4efa98620def0a08ca346e2a08209ab8d5 WatchSource:0}: Error finding container 94d11e220810737144e8281d689a2d4efa98620def0a08ca346e2a08209ab8d5: Status 404 returned error can't find the container with id 94d11e220810737144e8281d689a2d4efa98620def0a08ca346e2a08209ab8d5 Feb 27 19:39:41 crc kubenswrapper[4839]: I0227 19:39:41.902018 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" event={"ID":"99311d9f-5bd3-4280-a720-26be8b87b47f","Type":"ContainerStarted","Data":"5f7a29740ac71f63359619af66a2e5e3e775313bf5d4ff05652b9a6fbec7ac4b"} Feb 27 19:39:41 crc kubenswrapper[4839]: I0227 19:39:41.902346 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:41 crc kubenswrapper[4839]: I0227 19:39:41.902359 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" event={"ID":"99311d9f-5bd3-4280-a720-26be8b87b47f","Type":"ContainerStarted","Data":"94d11e220810737144e8281d689a2d4efa98620def0a08ca346e2a08209ab8d5"} Feb 27 19:39:41 crc kubenswrapper[4839]: I0227 19:39:41.906897 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" Feb 27 19:39:41 crc kubenswrapper[4839]: I0227 19:39:41.921607 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9645b9d-w9zp7" podStartSLOduration=67.921591602 podStartE2EDuration="1m7.921591602s" podCreationTimestamp="2026-02-27 19:38:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:39:41.919753623 +0000 UTC m=+343.564623358" watchObservedRunningTime="2026-02-27 19:39:41.921591602 +0000 UTC m=+343.566461337" Feb 27 19:39:43 crc kubenswrapper[4839]: E0227 19:39:43.967951 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:39:45 crc kubenswrapper[4839]: E0227 19:39:45.968428 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:39:52 crc kubenswrapper[4839]: E0227 19:39:52.968602 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:39:55 crc kubenswrapper[4839]: E0227 19:39:55.968117 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:39:59 crc kubenswrapper[4839]: E0227 19:39:58.995234 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.142384 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537020-pq8zb"] Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.143198 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.148423 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44lp8" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.148478 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.148581 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.154315 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537020-pq8zb"] Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.298446 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s7j5\" (UniqueName: \"kubernetes.io/projected/4fa05b9a-2335-4621-9c59-147a735f526f-kube-api-access-9s7j5\") pod \"auto-csr-approver-29537020-pq8zb\" (UID: \"4fa05b9a-2335-4621-9c59-147a735f526f\") " pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.399609 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s7j5\" (UniqueName: \"kubernetes.io/projected/4fa05b9a-2335-4621-9c59-147a735f526f-kube-api-access-9s7j5\") pod \"auto-csr-approver-29537020-pq8zb\" (UID: \"4fa05b9a-2335-4621-9c59-147a735f526f\") " pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.422465 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s7j5\" (UniqueName: \"kubernetes.io/projected/4fa05b9a-2335-4621-9c59-147a735f526f-kube-api-access-9s7j5\") pod \"auto-csr-approver-29537020-pq8zb\" (UID: \"4fa05b9a-2335-4621-9c59-147a735f526f\") " pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.503577 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:00 crc kubenswrapper[4839]: I0227 19:40:00.903219 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537020-pq8zb"] Feb 27 19:40:00 crc kubenswrapper[4839]: W0227 19:40:00.909427 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fa05b9a_2335_4621_9c59_147a735f526f.slice/crio-bb39290af7472a6726463f7634c239c32e406ef20a9ed8abd03aee340980371a WatchSource:0}: Error finding container bb39290af7472a6726463f7634c239c32e406ef20a9ed8abd03aee340980371a: Status 404 returned error can't find the container with id bb39290af7472a6726463f7634c239c32e406ef20a9ed8abd03aee340980371a Feb 27 19:40:01 crc kubenswrapper[4839]: I0227 19:40:01.032836 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" event={"ID":"4fa05b9a-2335-4621-9c59-147a735f526f","Type":"ContainerStarted","Data":"bb39290af7472a6726463f7634c239c32e406ef20a9ed8abd03aee340980371a"} Feb 27 19:40:02 crc kubenswrapper[4839]: E0227 19:40:02.364425 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:40:02 crc kubenswrapper[4839]: E0227 19:40:02.364974 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:40:02 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:40:02 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9s7j5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537020-pq8zb_openshift-infra(4fa05b9a-2335-4621-9c59-147a735f526f): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:40:02 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:40:02 crc kubenswrapper[4839]: E0227 19:40:02.366244 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" podUID="4fa05b9a-2335-4621-9c59-147a735f526f" Feb 27 19:40:03 crc kubenswrapper[4839]: E0227 19:40:03.046927 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" podUID="4fa05b9a-2335-4621-9c59-147a735f526f" Feb 27 19:40:03 crc kubenswrapper[4839]: E0227 19:40:03.966277 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:40:07 crc kubenswrapper[4839]: E0227 19:40:07.967952 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:40:10 crc kubenswrapper[4839]: E0227 19:40:10.967112 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.423596 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f2vhg"] Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.424906 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.443598 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f2vhg"] Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458632 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67653eda-ffa3-49cc-847c-6195dd5227e7-registry-certificates\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458731 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67653eda-ffa3-49cc-847c-6195dd5227e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458776 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwrxk\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-kube-api-access-gwrxk\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458800 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67653eda-ffa3-49cc-847c-6195dd5227e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458881 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-bound-sa-token\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458919 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-registry-tls\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458953 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.458984 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67653eda-ffa3-49cc-847c-6195dd5227e7-trusted-ca\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.484179 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.559801 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67653eda-ffa3-49cc-847c-6195dd5227e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.559954 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwrxk\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-kube-api-access-gwrxk\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.559998 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67653eda-ffa3-49cc-847c-6195dd5227e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.560037 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-bound-sa-token\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.560088 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-registry-tls\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.560126 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67653eda-ffa3-49cc-847c-6195dd5227e7-trusted-ca\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.560180 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67653eda-ffa3-49cc-847c-6195dd5227e7-registry-certificates\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.560454 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67653eda-ffa3-49cc-847c-6195dd5227e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.561584 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67653eda-ffa3-49cc-847c-6195dd5227e7-trusted-ca\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.562780 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67653eda-ffa3-49cc-847c-6195dd5227e7-registry-certificates\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.567227 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67653eda-ffa3-49cc-847c-6195dd5227e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.574248 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-registry-tls\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.579208 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwrxk\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-kube-api-access-gwrxk\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.587962 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67653eda-ffa3-49cc-847c-6195dd5227e7-bound-sa-token\") pod \"image-registry-66df7c8f76-f2vhg\" (UID: \"67653eda-ffa3-49cc-847c-6195dd5227e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:13 crc kubenswrapper[4839]: I0227 19:40:13.752038 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:14 crc kubenswrapper[4839]: I0227 19:40:14.232857 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f2vhg"] Feb 27 19:40:14 crc kubenswrapper[4839]: W0227 19:40:14.240112 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67653eda_ffa3_49cc_847c_6195dd5227e7.slice/crio-6ccb04b53ae3912d656473ffb746aa293cf1749b57b7c28cde7c1532b770a720 WatchSource:0}: Error finding container 6ccb04b53ae3912d656473ffb746aa293cf1749b57b7c28cde7c1532b770a720: Status 404 returned error can't find the container with id 6ccb04b53ae3912d656473ffb746aa293cf1749b57b7c28cde7c1532b770a720 Feb 27 19:40:15 crc kubenswrapper[4839]: I0227 19:40:15.127100 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" event={"ID":"67653eda-ffa3-49cc-847c-6195dd5227e7","Type":"ContainerStarted","Data":"69a69535f4338c1905a02aeffae7dfcb232ecc249f4c0fa04366d22340c82235"} Feb 27 19:40:15 crc kubenswrapper[4839]: I0227 19:40:15.127536 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" event={"ID":"67653eda-ffa3-49cc-847c-6195dd5227e7","Type":"ContainerStarted","Data":"6ccb04b53ae3912d656473ffb746aa293cf1749b57b7c28cde7c1532b770a720"} Feb 27 19:40:15 crc kubenswrapper[4839]: I0227 19:40:15.127563 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:16 crc kubenswrapper[4839]: I0227 19:40:16.599861 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" podStartSLOduration=3.599838295 podStartE2EDuration="3.599838295s" podCreationTimestamp="2026-02-27 19:40:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:40:15.149930875 +0000 UTC m=+376.794800650" watchObservedRunningTime="2026-02-27 19:40:16.599838295 +0000 UTC m=+378.244708040" Feb 27 19:40:16 crc kubenswrapper[4839]: I0227 19:40:16.603995 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5p9bn"] Feb 27 19:40:16 crc kubenswrapper[4839]: I0227 19:40:16.604303 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5p9bn" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="registry-server" containerID="cri-o://6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3" gracePeriod=2 Feb 27 19:40:16 crc kubenswrapper[4839]: E0227 19:40:16.966937 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:40:16 crc kubenswrapper[4839]: I0227 19:40:16.994987 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.045464 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvsn8\" (UniqueName: \"kubernetes.io/projected/8cc72c87-b6d6-4384-ac6e-fb12af841f66-kube-api-access-mvsn8\") pod \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.045510 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-utilities\") pod \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.045539 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-catalog-content\") pod \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\" (UID: \"8cc72c87-b6d6-4384-ac6e-fb12af841f66\") " Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.046561 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-utilities" (OuterVolumeSpecName: "utilities") pod "8cc72c87-b6d6-4384-ac6e-fb12af841f66" (UID: "8cc72c87-b6d6-4384-ac6e-fb12af841f66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.051790 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc72c87-b6d6-4384-ac6e-fb12af841f66-kube-api-access-mvsn8" (OuterVolumeSpecName: "kube-api-access-mvsn8") pod "8cc72c87-b6d6-4384-ac6e-fb12af841f66" (UID: "8cc72c87-b6d6-4384-ac6e-fb12af841f66"). InnerVolumeSpecName "kube-api-access-mvsn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.094129 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cc72c87-b6d6-4384-ac6e-fb12af841f66" (UID: "8cc72c87-b6d6-4384-ac6e-fb12af841f66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.141699 4839 generic.go:334] "Generic (PLEG): container finished" podID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerID="6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3" exitCode=0 Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.141755 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5p9bn" event={"ID":"8cc72c87-b6d6-4384-ac6e-fb12af841f66","Type":"ContainerDied","Data":"6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3"} Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.141796 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5p9bn" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.142049 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5p9bn" event={"ID":"8cc72c87-b6d6-4384-ac6e-fb12af841f66","Type":"ContainerDied","Data":"34950ce8717a42cbebef568b404118724c4474152824aed30cea99a537383ff9"} Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.142075 4839 scope.go:117] "RemoveContainer" containerID="6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.146988 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvsn8\" (UniqueName: \"kubernetes.io/projected/8cc72c87-b6d6-4384-ac6e-fb12af841f66-kube-api-access-mvsn8\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.147215 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.147314 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc72c87-b6d6-4384-ac6e-fb12af841f66-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.156812 4839 scope.go:117] "RemoveContainer" containerID="d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.171626 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5p9bn"] Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.175008 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5p9bn"] Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.202419 4839 scope.go:117] "RemoveContainer" containerID="2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.218867 4839 scope.go:117] "RemoveContainer" containerID="6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3" Feb 27 19:40:17 crc kubenswrapper[4839]: E0227 19:40:17.219182 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3\": container with ID starting with 6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3 not found: ID does not exist" containerID="6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.219226 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3"} err="failed to get container status \"6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3\": rpc error: code = NotFound desc = could not find container \"6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3\": container with ID starting with 6de32cb98921cb80a425e2a0dd77d80e8a91c387fe90bf220adb5a7cb4979fc3 not found: ID does not exist" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.219252 4839 scope.go:117] "RemoveContainer" containerID="d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2" Feb 27 19:40:17 crc kubenswrapper[4839]: E0227 19:40:17.219682 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2\": container with ID starting with d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2 not found: ID does not exist" containerID="d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.219752 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2"} err="failed to get container status \"d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2\": rpc error: code = NotFound desc = could not find container \"d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2\": container with ID starting with d2ddeb94e214e6f742cfcfa2b5803a20d586597466fa8488b341aff72baf94f2 not found: ID does not exist" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.219795 4839 scope.go:117] "RemoveContainer" containerID="2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c" Feb 27 19:40:17 crc kubenswrapper[4839]: E0227 19:40:17.220099 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c\": container with ID starting with 2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c not found: ID does not exist" containerID="2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c" Feb 27 19:40:17 crc kubenswrapper[4839]: I0227 19:40:17.220140 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c"} err="failed to get container status \"2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c\": rpc error: code = NotFound desc = could not find container \"2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c\": container with ID starting with 2384606ab90b7ebec912d1aaad25efed673ad5455ba6a1f1ea7cc55f92f2339c not found: ID does not exist" Feb 27 19:40:18 crc kubenswrapper[4839]: E0227 19:40:18.970102 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" Feb 27 19:40:18 crc kubenswrapper[4839]: I0227 19:40:18.972477 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" path="/var/lib/kubelet/pods/8cc72c87-b6d6-4384-ac6e-fb12af841f66/volumes" Feb 27 19:40:19 crc kubenswrapper[4839]: I0227 19:40:19.154588 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" event={"ID":"4fa05b9a-2335-4621-9c59-147a735f526f","Type":"ContainerStarted","Data":"ff14ddc375cad7acab149b6cd2cedd080fbdeb204685691990524468298d4134"} Feb 27 19:40:19 crc kubenswrapper[4839]: I0227 19:40:19.170245 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" podStartSLOduration=1.232204539 podStartE2EDuration="19.170232109s" podCreationTimestamp="2026-02-27 19:40:00 +0000 UTC" firstStartedPulling="2026-02-27 19:40:00.911865185 +0000 UTC m=+362.556734920" lastFinishedPulling="2026-02-27 19:40:18.849892715 +0000 UTC m=+380.494762490" observedRunningTime="2026-02-27 19:40:19.167119745 +0000 UTC m=+380.811989480" watchObservedRunningTime="2026-02-27 19:40:19.170232109 +0000 UTC m=+380.815101844" Feb 27 19:40:20 crc kubenswrapper[4839]: I0227 19:40:20.161974 4839 generic.go:334] "Generic (PLEG): container finished" podID="4fa05b9a-2335-4621-9c59-147a735f526f" containerID="ff14ddc375cad7acab149b6cd2cedd080fbdeb204685691990524468298d4134" exitCode=0 Feb 27 19:40:20 crc kubenswrapper[4839]: I0227 19:40:20.162023 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" event={"ID":"4fa05b9a-2335-4621-9c59-147a735f526f","Type":"ContainerDied","Data":"ff14ddc375cad7acab149b6cd2cedd080fbdeb204685691990524468298d4134"} Feb 27 19:40:21 crc kubenswrapper[4839]: I0227 19:40:21.454450 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:21 crc kubenswrapper[4839]: I0227 19:40:21.506399 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s7j5\" (UniqueName: \"kubernetes.io/projected/4fa05b9a-2335-4621-9c59-147a735f526f-kube-api-access-9s7j5\") pod \"4fa05b9a-2335-4621-9c59-147a735f526f\" (UID: \"4fa05b9a-2335-4621-9c59-147a735f526f\") " Feb 27 19:40:21 crc kubenswrapper[4839]: I0227 19:40:21.513237 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa05b9a-2335-4621-9c59-147a735f526f-kube-api-access-9s7j5" (OuterVolumeSpecName: "kube-api-access-9s7j5") pod "4fa05b9a-2335-4621-9c59-147a735f526f" (UID: "4fa05b9a-2335-4621-9c59-147a735f526f"). InnerVolumeSpecName "kube-api-access-9s7j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:40:21 crc kubenswrapper[4839]: I0227 19:40:21.608542 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s7j5\" (UniqueName: \"kubernetes.io/projected/4fa05b9a-2335-4621-9c59-147a735f526f-kube-api-access-9s7j5\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:22 crc kubenswrapper[4839]: I0227 19:40:22.189056 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" event={"ID":"4fa05b9a-2335-4621-9c59-147a735f526f","Type":"ContainerDied","Data":"bb39290af7472a6726463f7634c239c32e406ef20a9ed8abd03aee340980371a"} Feb 27 19:40:22 crc kubenswrapper[4839]: I0227 19:40:22.189354 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb39290af7472a6726463f7634c239c32e406ef20a9ed8abd03aee340980371a" Feb 27 19:40:22 crc kubenswrapper[4839]: I0227 19:40:22.189137 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537020-pq8zb" Feb 27 19:40:23 crc kubenswrapper[4839]: E0227 19:40:23.967849 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" Feb 27 19:40:31 crc kubenswrapper[4839]: E0227 19:40:31.540865 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:40:31 crc kubenswrapper[4839]: E0227 19:40:31.541505 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n76bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ff888_openshift-marketplace(68a8a402-5d27-4632-8cf2-aaccccfff95d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:40:31 crc kubenswrapper[4839]: E0227 19:40:31.542750 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:40:33 crc kubenswrapper[4839]: I0227 19:40:33.759989 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-f2vhg" Feb 27 19:40:33 crc kubenswrapper[4839]: I0227 19:40:33.819435 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf6mj"] Feb 27 19:40:37 crc kubenswrapper[4839]: I0227 19:40:37.293332 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerStarted","Data":"c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf"} Feb 27 19:40:38 crc kubenswrapper[4839]: I0227 19:40:38.300828 4839 generic.go:334] "Generic (PLEG): container finished" podID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerID="c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf" exitCode=0 Feb 27 19:40:38 crc kubenswrapper[4839]: I0227 19:40:38.300904 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerDied","Data":"c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf"} Feb 27 19:40:39 crc kubenswrapper[4839]: I0227 19:40:39.307895 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerStarted","Data":"9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae"} Feb 27 19:40:39 crc kubenswrapper[4839]: I0227 19:40:39.327797 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgzcw" podStartSLOduration=2.401614946 podStartE2EDuration="3m7.327777886s" podCreationTimestamp="2026-02-27 19:37:32 +0000 UTC" firstStartedPulling="2026-02-27 19:37:33.855463239 +0000 UTC m=+215.500332974" lastFinishedPulling="2026-02-27 19:40:38.781626169 +0000 UTC m=+400.426495914" observedRunningTime="2026-02-27 19:40:39.325778964 +0000 UTC m=+400.970648709" watchObservedRunningTime="2026-02-27 19:40:39.327777886 +0000 UTC m=+400.972647621" Feb 27 19:40:40 crc kubenswrapper[4839]: I0227 19:40:40.314423 4839 generic.go:334] "Generic (PLEG): container finished" podID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerID="3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad" exitCode=0 Feb 27 19:40:40 crc kubenswrapper[4839]: I0227 19:40:40.314475 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2p75" event={"ID":"b471c70a-15fe-4b5f-aa6f-354848925ace","Type":"ContainerDied","Data":"3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad"} Feb 27 19:40:41 crc kubenswrapper[4839]: I0227 19:40:41.325268 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2p75" event={"ID":"b471c70a-15fe-4b5f-aa6f-354848925ace","Type":"ContainerStarted","Data":"5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb"} Feb 27 19:40:41 crc kubenswrapper[4839]: I0227 19:40:41.357448 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d2p75" podStartSLOduration=3.464189849 podStartE2EDuration="3m10.357426965s" podCreationTimestamp="2026-02-27 19:37:31 +0000 UTC" firstStartedPulling="2026-02-27 19:37:33.82963598 +0000 UTC m=+215.474505715" lastFinishedPulling="2026-02-27 19:40:40.722873096 +0000 UTC m=+402.367742831" observedRunningTime="2026-02-27 19:40:41.351569092 +0000 UTC m=+402.996438837" watchObservedRunningTime="2026-02-27 19:40:41.357426965 +0000 UTC m=+403.002296720" Feb 27 19:40:42 crc kubenswrapper[4839]: I0227 19:40:42.260909 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:40:42 crc kubenswrapper[4839]: I0227 19:40:42.260964 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:40:42 crc kubenswrapper[4839]: I0227 19:40:42.595510 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:40:42 crc kubenswrapper[4839]: I0227 19:40:42.595587 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:40:42 crc kubenswrapper[4839]: I0227 19:40:42.647430 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:40:43 crc kubenswrapper[4839]: I0227 19:40:43.294220 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="registry-server" probeResult="failure" output=< Feb 27 19:40:43 crc kubenswrapper[4839]: timeout: failed to connect service ":50051" within 1s Feb 27 19:40:43 crc kubenswrapper[4839]: > Feb 27 19:40:43 crc kubenswrapper[4839]: I0227 19:40:43.384012 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:40:43 crc kubenswrapper[4839]: E0227 19:40:43.968769 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:40:44 crc kubenswrapper[4839]: I0227 19:40:44.805143 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgzcw"] Feb 27 19:40:45 crc kubenswrapper[4839]: I0227 19:40:45.349443 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgzcw" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="registry-server" containerID="cri-o://9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae" gracePeriod=2 Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.242873 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.322810 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-catalog-content\") pod \"0a4d0578-88dd-4754-81fd-ae28a709b698\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.322917 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-utilities\") pod \"0a4d0578-88dd-4754-81fd-ae28a709b698\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.322979 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm9fn\" (UniqueName: \"kubernetes.io/projected/0a4d0578-88dd-4754-81fd-ae28a709b698-kube-api-access-zm9fn\") pod \"0a4d0578-88dd-4754-81fd-ae28a709b698\" (UID: \"0a4d0578-88dd-4754-81fd-ae28a709b698\") " Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.324906 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-utilities" (OuterVolumeSpecName: "utilities") pod "0a4d0578-88dd-4754-81fd-ae28a709b698" (UID: "0a4d0578-88dd-4754-81fd-ae28a709b698"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.331114 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4d0578-88dd-4754-81fd-ae28a709b698-kube-api-access-zm9fn" (OuterVolumeSpecName: "kube-api-access-zm9fn") pod "0a4d0578-88dd-4754-81fd-ae28a709b698" (UID: "0a4d0578-88dd-4754-81fd-ae28a709b698"). InnerVolumeSpecName "kube-api-access-zm9fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.356762 4839 generic.go:334] "Generic (PLEG): container finished" podID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerID="9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae" exitCode=0 Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.356832 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerDied","Data":"9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae"} Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.356851 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgzcw" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.356861 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgzcw" event={"ID":"0a4d0578-88dd-4754-81fd-ae28a709b698","Type":"ContainerDied","Data":"bc1ef7693019441d1a87decbad2ec0166bf4bf3d76d5293cf63f3e4acd0fd5cf"} Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.356879 4839 scope.go:117] "RemoveContainer" containerID="9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.376596 4839 scope.go:117] "RemoveContainer" containerID="c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.393188 4839 scope.go:117] "RemoveContainer" containerID="4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.408234 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a4d0578-88dd-4754-81fd-ae28a709b698" (UID: "0a4d0578-88dd-4754-81fd-ae28a709b698"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.409634 4839 scope.go:117] "RemoveContainer" containerID="9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae" Feb 27 19:40:46 crc kubenswrapper[4839]: E0227 19:40:46.410225 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae\": container with ID starting with 9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae not found: ID does not exist" containerID="9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.410261 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae"} err="failed to get container status \"9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae\": rpc error: code = NotFound desc = could not find container \"9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae\": container with ID starting with 9c395e03c1ca09bc92baf1467fc65139ed1029fd2755471da7812b55ef8440ae not found: ID does not exist" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.410282 4839 scope.go:117] "RemoveContainer" containerID="c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf" Feb 27 19:40:46 crc kubenswrapper[4839]: E0227 19:40:46.410777 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf\": container with ID starting with c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf not found: ID does not exist" containerID="c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.410802 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf"} err="failed to get container status \"c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf\": rpc error: code = NotFound desc = could not find container \"c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf\": container with ID starting with c4d3b2b516e2ee71e74564cc9c56a30953f5dc7558314de7b56315a5cc522bcf not found: ID does not exist" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.410820 4839 scope.go:117] "RemoveContainer" containerID="4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1" Feb 27 19:40:46 crc kubenswrapper[4839]: E0227 19:40:46.411279 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1\": container with ID starting with 4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1 not found: ID does not exist" containerID="4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.411326 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1"} err="failed to get container status \"4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1\": rpc error: code = NotFound desc = could not find container \"4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1\": container with ID starting with 4b627b08e237d742c36d630d0e05edd1d6ed22fd1037028b51df278782789cd1 not found: ID does not exist" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.424509 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm9fn\" (UniqueName: \"kubernetes.io/projected/0a4d0578-88dd-4754-81fd-ae28a709b698-kube-api-access-zm9fn\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.424540 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.424553 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d0578-88dd-4754-81fd-ae28a709b698-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.683408 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgzcw"] Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.690314 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgzcw"] Feb 27 19:40:46 crc kubenswrapper[4839]: I0227 19:40:46.973453 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" path="/var/lib/kubelet/pods/0a4d0578-88dd-4754-81fd-ae28a709b698/volumes" Feb 27 19:40:52 crc kubenswrapper[4839]: I0227 19:40:52.309938 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:40:52 crc kubenswrapper[4839]: I0227 19:40:52.671783 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:40:52 crc kubenswrapper[4839]: I0227 19:40:52.682179 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:40:52 crc kubenswrapper[4839]: I0227 19:40:52.682259 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:40:57 crc kubenswrapper[4839]: E0227 19:40:57.968062 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:40:58 crc kubenswrapper[4839]: I0227 19:40:58.878571 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" podUID="e4bdb150-3f14-4eac-9f05-e10a515c7cb4" containerName="registry" containerID="cri-o://533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57" gracePeriod=30 Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.323425 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.436459 4839 generic.go:334] "Generic (PLEG): container finished" podID="e4bdb150-3f14-4eac-9f05-e10a515c7cb4" containerID="533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57" exitCode=0 Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.436500 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" event={"ID":"e4bdb150-3f14-4eac-9f05-e10a515c7cb4","Type":"ContainerDied","Data":"533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57"} Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.436533 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" event={"ID":"e4bdb150-3f14-4eac-9f05-e10a515c7cb4","Type":"ContainerDied","Data":"6aaa1982c32bed4aa521079631d6e92aacc86204cdce13b3f400e645bc76e15a"} Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.436532 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf6mj" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.436549 4839 scope.go:117] "RemoveContainer" containerID="533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.456786 4839 scope.go:117] "RemoveContainer" containerID="533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57" Feb 27 19:40:59 crc kubenswrapper[4839]: E0227 19:40:59.457276 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57\": container with ID starting with 533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57 not found: ID does not exist" containerID="533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.457325 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57"} err="failed to get container status \"533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57\": rpc error: code = NotFound desc = could not find container \"533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57\": container with ID starting with 533bcadbc6d03d447b8bc81b39a5588cdef61f7306e840bf6594ac58c8198f57 not found: ID does not exist" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501148 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501203 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-installation-pull-secrets\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501239 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-certificates\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501284 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-ca-trust-extracted\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501315 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp2zk\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-kube-api-access-cp2zk\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501480 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-tls\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501524 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-trusted-ca\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501577 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-bound-sa-token\") pod \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\" (UID: \"e4bdb150-3f14-4eac-9f05-e10a515c7cb4\") " Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.501908 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.502427 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.507382 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.507395 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.508227 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.510614 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.514874 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-kube-api-access-cp2zk" (OuterVolumeSpecName: "kube-api-access-cp2zk") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "kube-api-access-cp2zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.523979 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e4bdb150-3f14-4eac-9f05-e10a515c7cb4" (UID: "e4bdb150-3f14-4eac-9f05-e10a515c7cb4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603402 4839 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603466 4839 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603487 4839 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603509 4839 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603528 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp2zk\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-kube-api-access-cp2zk\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603545 4839 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.603561 4839 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4bdb150-3f14-4eac-9f05-e10a515c7cb4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.779963 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf6mj"] Feb 27 19:40:59 crc kubenswrapper[4839]: I0227 19:40:59.784471 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf6mj"] Feb 27 19:41:00 crc kubenswrapper[4839]: I0227 19:41:00.977334 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4bdb150-3f14-4eac-9f05-e10a515c7cb4" path="/var/lib/kubelet/pods/e4bdb150-3f14-4eac-9f05-e10a515c7cb4/volumes" Feb 27 19:41:08 crc kubenswrapper[4839]: E0227 19:41:08.971820 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ff888" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.051650 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-stz5c"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.051911 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-stz5c" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="registry-server" containerID="cri-o://c062e48e1dcfbe0524b8bf04be3045ea174c4a5b81b1200ae146399c4db1bc29" gracePeriod=30 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.062934 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d2p75"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.063193 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d2p75" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="registry-server" containerID="cri-o://5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb" gracePeriod=30 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.068829 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-clhjs"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.069040 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerName="marketplace-operator" containerID="cri-o://a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372" gracePeriod=30 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.072409 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fddl"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.073243 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7fddl" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="registry-server" containerID="cri-o://f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1" gracePeriod=30 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.087602 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff888"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.094818 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gdwpn"] Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095126 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="registry-server" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.095226 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="registry-server" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095306 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="extract-content" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.095374 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="extract-content" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095435 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="extract-utilities" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.095498 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="extract-utilities" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095560 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="extract-content" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.095627 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="extract-content" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095707 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="registry-server" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.095778 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="registry-server" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095844 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="extract-utilities" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.095908 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="extract-utilities" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.095965 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa05b9a-2335-4621-9c59-147a735f526f" containerName="oc" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.096022 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa05b9a-2335-4621-9c59-147a735f526f" containerName="oc" Feb 27 19:41:09 crc kubenswrapper[4839]: E0227 19:41:09.096085 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bdb150-3f14-4eac-9f05-e10a515c7cb4" containerName="registry" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.096146 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bdb150-3f14-4eac-9f05-e10a515c7cb4" containerName="registry" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.096315 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc72c87-b6d6-4384-ac6e-fb12af841f66" containerName="registry-server" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.096386 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4d0578-88dd-4754-81fd-ae28a709b698" containerName="registry-server" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.096452 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4bdb150-3f14-4eac-9f05-e10a515c7cb4" containerName="registry" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.096515 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa05b9a-2335-4621-9c59-147a735f526f" containerName="oc" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.097009 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.101295 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m6qmh"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.102000 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m6qmh" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="registry-server" containerID="cri-o://13d27d812bf2e04ba7894fd3d40f834e035ea6cc1d6d6bd4adb237611ec7d7c4" gracePeriod=30 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.116057 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gdwpn"] Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.227528 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.227801 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cskzw\" (UniqueName: \"kubernetes.io/projected/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-kube-api-access-cskzw\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.227823 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.328820 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.328937 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.328976 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cskzw\" (UniqueName: \"kubernetes.io/projected/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-kube-api-access-cskzw\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.331072 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.336958 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.349281 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cskzw\" (UniqueName: \"kubernetes.io/projected/ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7-kube-api-access-cskzw\") pod \"marketplace-operator-79b997595-gdwpn\" (UID: \"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.462710 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.469344 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.501948 4839 generic.go:334] "Generic (PLEG): container finished" podID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerID="5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb" exitCode=0 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.502031 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2p75" event={"ID":"b471c70a-15fe-4b5f-aa6f-354848925ace","Type":"ContainerDied","Data":"5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.502061 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2p75" event={"ID":"b471c70a-15fe-4b5f-aa6f-354848925ace","Type":"ContainerDied","Data":"dd897366779f7f72927bd14fe45246f0517568162e089987e03c69a68e6e97ac"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.502084 4839 scope.go:117] "RemoveContainer" containerID="5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.502318 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2p75" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.504108 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.505786 4839 generic.go:334] "Generic (PLEG): container finished" podID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerID="a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372" exitCode=0 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.505879 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" event={"ID":"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2","Type":"ContainerDied","Data":"a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.505942 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" event={"ID":"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2","Type":"ContainerDied","Data":"d2395d253a18500bee67e573a7a2fb68bcaee9c7a072b88316100bf5fa0d0907"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.509464 4839 generic.go:334] "Generic (PLEG): container finished" podID="f11b8271-1de0-432d-8d78-2379862726a1" containerID="13d27d812bf2e04ba7894fd3d40f834e035ea6cc1d6d6bd4adb237611ec7d7c4" exitCode=0 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.509515 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6qmh" event={"ID":"f11b8271-1de0-432d-8d78-2379862726a1","Type":"ContainerDied","Data":"13d27d812bf2e04ba7894fd3d40f834e035ea6cc1d6d6bd4adb237611ec7d7c4"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.512640 4839 generic.go:334] "Generic (PLEG): container finished" podID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerID="c062e48e1dcfbe0524b8bf04be3045ea174c4a5b81b1200ae146399c4db1bc29" exitCode=0 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.513416 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-stz5c" event={"ID":"ff0a141c-3800-4c84-b1d6-17ec909d5d27","Type":"ContainerDied","Data":"c062e48e1dcfbe0524b8bf04be3045ea174c4a5b81b1200ae146399c4db1bc29"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.513751 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-stz5c" event={"ID":"ff0a141c-3800-4c84-b1d6-17ec909d5d27","Type":"ContainerDied","Data":"1b1f044e4baad409b2fd46ee476b0abf1db6187b6885076a6f23f0d06edcaa21"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.513904 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b1f044e4baad409b2fd46ee476b0abf1db6187b6885076a6f23f0d06edcaa21" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.518284 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.519771 4839 scope.go:117] "RemoveContainer" containerID="3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.529610 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.530617 4839 generic.go:334] "Generic (PLEG): container finished" podID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerID="f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1" exitCode=0 Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.530789 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerDied","Data":"f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.530812 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fddl" event={"ID":"67e667b9-304d-4052-8f0f-8d46d920aa7c","Type":"ContainerDied","Data":"d57f6811ea530e069dd3c348bfdcb5b9cf37d47bca3b133b5442cc82ecc62964"} Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.531512 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:41:09 crc kubenswrapper[4839]: I0227 19:41:09.554962 4839 scope.go:117] "RemoveContainer" containerID="89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.603174 4839 scope.go:117] "RemoveContainer" containerID="5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.603590 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb\": container with ID starting with 5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb not found: ID does not exist" containerID="5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.603632 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb"} err="failed to get container status \"5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb\": rpc error: code = NotFound desc = could not find container \"5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb\": container with ID starting with 5b43645a0748ebc22ad42edf53586b5e5befc5f806bcc3b033e72fb116500dfb not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.603675 4839 scope.go:117] "RemoveContainer" containerID="3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.603953 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad\": container with ID starting with 3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad not found: ID does not exist" containerID="3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.603976 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad"} err="failed to get container status \"3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad\": rpc error: code = NotFound desc = could not find container \"3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad\": container with ID starting with 3991299389f395da260c6751d53ac1204e57d0d73fd6afd8eb4b2105812393ad not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.603991 4839 scope.go:117] "RemoveContainer" containerID="89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.604430 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab\": container with ID starting with 89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab not found: ID does not exist" containerID="89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.604478 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab"} err="failed to get container status \"89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab\": rpc error: code = NotFound desc = could not find container \"89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab\": container with ID starting with 89395b9f11ca305356892d81741de41ed23663cc9a439f6d082cdf32e830a8ab not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.604511 4839 scope.go:117] "RemoveContainer" containerID="a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.615892 4839 scope.go:117] "RemoveContainer" containerID="a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.616322 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372\": container with ID starting with a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372 not found: ID does not exist" containerID="a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.616352 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372"} err="failed to get container status \"a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372\": rpc error: code = NotFound desc = could not find container \"a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372\": container with ID starting with a9cf7d357f08ae4e71b7ce1291652a713d89e87a62f91fbde6e2d58b1913c372 not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.616373 4839 scope.go:117] "RemoveContainer" containerID="f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.629964 4839 scope.go:117] "RemoveContainer" containerID="740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.634894 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-catalog-content\") pod \"b471c70a-15fe-4b5f-aa6f-354848925ace\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.634940 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-trusted-ca\") pod \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.634973 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-catalog-content\") pod \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635004 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-catalog-content\") pod \"67e667b9-304d-4052-8f0f-8d46d920aa7c\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635023 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8d85\" (UniqueName: \"kubernetes.io/projected/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-kube-api-access-k8d85\") pod \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635038 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-catalog-content\") pod \"f11b8271-1de0-432d-8d78-2379862726a1\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635061 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w5zs\" (UniqueName: \"kubernetes.io/projected/67e667b9-304d-4052-8f0f-8d46d920aa7c-kube-api-access-8w5zs\") pod \"67e667b9-304d-4052-8f0f-8d46d920aa7c\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635080 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-operator-metrics\") pod \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\" (UID: \"2cb7672b-b64f-4e33-9ea3-939a2a6e9da2\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635095 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlf2r\" (UniqueName: \"kubernetes.io/projected/f11b8271-1de0-432d-8d78-2379862726a1-kube-api-access-nlf2r\") pod \"f11b8271-1de0-432d-8d78-2379862726a1\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635123 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-utilities\") pod \"b471c70a-15fe-4b5f-aa6f-354848925ace\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635152 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8btrd\" (UniqueName: \"kubernetes.io/projected/ff0a141c-3800-4c84-b1d6-17ec909d5d27-kube-api-access-8btrd\") pod \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635171 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-utilities\") pod \"67e667b9-304d-4052-8f0f-8d46d920aa7c\" (UID: \"67e667b9-304d-4052-8f0f-8d46d920aa7c\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635200 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-utilities\") pod \"f11b8271-1de0-432d-8d78-2379862726a1\" (UID: \"f11b8271-1de0-432d-8d78-2379862726a1\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635233 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5wrd\" (UniqueName: \"kubernetes.io/projected/b471c70a-15fe-4b5f-aa6f-354848925ace-kube-api-access-z5wrd\") pod \"b471c70a-15fe-4b5f-aa6f-354848925ace\" (UID: \"b471c70a-15fe-4b5f-aa6f-354848925ace\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.635279 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-utilities\") pod \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\" (UID: \"ff0a141c-3800-4c84-b1d6-17ec909d5d27\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.636522 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-utilities" (OuterVolumeSpecName: "utilities") pod "67e667b9-304d-4052-8f0f-8d46d920aa7c" (UID: "67e667b9-304d-4052-8f0f-8d46d920aa7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.636652 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-utilities" (OuterVolumeSpecName: "utilities") pod "ff0a141c-3800-4c84-b1d6-17ec909d5d27" (UID: "ff0a141c-3800-4c84-b1d6-17ec909d5d27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.636740 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-utilities" (OuterVolumeSpecName: "utilities") pod "b471c70a-15fe-4b5f-aa6f-354848925ace" (UID: "b471c70a-15fe-4b5f-aa6f-354848925ace"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.638535 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-utilities" (OuterVolumeSpecName: "utilities") pod "f11b8271-1de0-432d-8d78-2379862726a1" (UID: "f11b8271-1de0-432d-8d78-2379862726a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.640915 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b471c70a-15fe-4b5f-aa6f-354848925ace-kube-api-access-z5wrd" (OuterVolumeSpecName: "kube-api-access-z5wrd") pod "b471c70a-15fe-4b5f-aa6f-354848925ace" (UID: "b471c70a-15fe-4b5f-aa6f-354848925ace"). InnerVolumeSpecName "kube-api-access-z5wrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.641084 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0a141c-3800-4c84-b1d6-17ec909d5d27-kube-api-access-8btrd" (OuterVolumeSpecName: "kube-api-access-8btrd") pod "ff0a141c-3800-4c84-b1d6-17ec909d5d27" (UID: "ff0a141c-3800-4c84-b1d6-17ec909d5d27"). InnerVolumeSpecName "kube-api-access-8btrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.641164 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e667b9-304d-4052-8f0f-8d46d920aa7c-kube-api-access-8w5zs" (OuterVolumeSpecName: "kube-api-access-8w5zs") pod "67e667b9-304d-4052-8f0f-8d46d920aa7c" (UID: "67e667b9-304d-4052-8f0f-8d46d920aa7c"). InnerVolumeSpecName "kube-api-access-8w5zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.641294 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-kube-api-access-k8d85" (OuterVolumeSpecName: "kube-api-access-k8d85") pod "2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" (UID: "2cb7672b-b64f-4e33-9ea3-939a2a6e9da2"). InnerVolumeSpecName "kube-api-access-k8d85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.641405 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" (UID: "2cb7672b-b64f-4e33-9ea3-939a2a6e9da2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.641615 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f11b8271-1de0-432d-8d78-2379862726a1-kube-api-access-nlf2r" (OuterVolumeSpecName: "kube-api-access-nlf2r") pod "f11b8271-1de0-432d-8d78-2379862726a1" (UID: "f11b8271-1de0-432d-8d78-2379862726a1"). InnerVolumeSpecName "kube-api-access-nlf2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.650281 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" (UID: "2cb7672b-b64f-4e33-9ea3-939a2a6e9da2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.657079 4839 scope.go:117] "RemoveContainer" containerID="bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.677931 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67e667b9-304d-4052-8f0f-8d46d920aa7c" (UID: "67e667b9-304d-4052-8f0f-8d46d920aa7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.683605 4839 scope.go:117] "RemoveContainer" containerID="f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.686826 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1\": container with ID starting with f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1 not found: ID does not exist" containerID="f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.686857 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1"} err="failed to get container status \"f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1\": rpc error: code = NotFound desc = could not find container \"f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1\": container with ID starting with f5c8f1b51d45ebbcbec89dfbc60679fbbe57d1d765923fadab499d261e76f1f1 not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.686886 4839 scope.go:117] "RemoveContainer" containerID="740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.687218 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c\": container with ID starting with 740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c not found: ID does not exist" containerID="740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.687234 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c"} err="failed to get container status \"740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c\": rpc error: code = NotFound desc = could not find container \"740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c\": container with ID starting with 740189a7f3284d25acd01ec4653d7b2e3b462904a106c1f25aa9dde3ed9bcd9c not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.687245 4839 scope.go:117] "RemoveContainer" containerID="bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b" Feb 27 19:41:10 crc kubenswrapper[4839]: E0227 19:41:09.687548 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b\": container with ID starting with bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b not found: ID does not exist" containerID="bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.687563 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b"} err="failed to get container status \"bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b\": rpc error: code = NotFound desc = could not find container \"bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b\": container with ID starting with bf6d52e431238779d730e841bdd8d88dded97ce204c2eee621b41e2b1697780b not found: ID does not exist" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.698979 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff0a141c-3800-4c84-b1d6-17ec909d5d27" (UID: "ff0a141c-3800-4c84-b1d6-17ec909d5d27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.699655 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b471c70a-15fe-4b5f-aa6f-354848925ace" (UID: "b471c70a-15fe-4b5f-aa6f-354848925ace"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736755 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736789 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736806 4839 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736819 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0a141c-3800-4c84-b1d6-17ec909d5d27-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736830 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736842 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8d85\" (UniqueName: \"kubernetes.io/projected/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-kube-api-access-k8d85\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736852 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w5zs\" (UniqueName: \"kubernetes.io/projected/67e667b9-304d-4052-8f0f-8d46d920aa7c-kube-api-access-8w5zs\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736863 4839 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736874 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlf2r\" (UniqueName: \"kubernetes.io/projected/f11b8271-1de0-432d-8d78-2379862726a1-kube-api-access-nlf2r\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736886 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b471c70a-15fe-4b5f-aa6f-354848925ace-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736897 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8btrd\" (UniqueName: \"kubernetes.io/projected/ff0a141c-3800-4c84-b1d6-17ec909d5d27-kube-api-access-8btrd\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736907 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67e667b9-304d-4052-8f0f-8d46d920aa7c-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736917 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.736927 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5wrd\" (UniqueName: \"kubernetes.io/projected/b471c70a-15fe-4b5f-aa6f-354848925ace-kube-api-access-z5wrd\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.770274 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f11b8271-1de0-432d-8d78-2379862726a1" (UID: "f11b8271-1de0-432d-8d78-2379862726a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.838091 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11b8271-1de0-432d-8d78-2379862726a1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.838436 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d2p75"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:09.842722 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d2p75"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.424304 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.537654 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ff888" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.537695 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ff888" event={"ID":"68a8a402-5d27-4632-8cf2-aaccccfff95d","Type":"ContainerDied","Data":"2bfa82e174ea9fbabffe183a7d7d3dd3b62d4d47add2072e2b7a770b63adc1e1"} Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.538445 4839 scope.go:117] "RemoveContainer" containerID="f8176690a7343e0e78625f76fa3e014e129765fac6f6dac057970c1bced3a2a6" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.539943 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-clhjs" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.541793 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6qmh" event={"ID":"f11b8271-1de0-432d-8d78-2379862726a1","Type":"ContainerDied","Data":"1677b5828c5ec3c5f293826a708719dbfce76bc862659d0abe096d491698c334"} Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.541912 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m6qmh" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.543781 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-stz5c" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.544796 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fddl" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.547117 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n76bg\" (UniqueName: \"kubernetes.io/projected/68a8a402-5d27-4632-8cf2-aaccccfff95d-kube-api-access-n76bg\") pod \"68a8a402-5d27-4632-8cf2-aaccccfff95d\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.547227 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-utilities\") pod \"68a8a402-5d27-4632-8cf2-aaccccfff95d\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.547256 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-catalog-content\") pod \"68a8a402-5d27-4632-8cf2-aaccccfff95d\" (UID: \"68a8a402-5d27-4632-8cf2-aaccccfff95d\") " Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.547802 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68a8a402-5d27-4632-8cf2-aaccccfff95d" (UID: "68a8a402-5d27-4632-8cf2-aaccccfff95d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.550774 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a8a402-5d27-4632-8cf2-aaccccfff95d-kube-api-access-n76bg" (OuterVolumeSpecName: "kube-api-access-n76bg") pod "68a8a402-5d27-4632-8cf2-aaccccfff95d" (UID: "68a8a402-5d27-4632-8cf2-aaccccfff95d"). InnerVolumeSpecName "kube-api-access-n76bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.551372 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-utilities" (OuterVolumeSpecName: "utilities") pod "68a8a402-5d27-4632-8cf2-aaccccfff95d" (UID: "68a8a402-5d27-4632-8cf2-aaccccfff95d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.564150 4839 scope.go:117] "RemoveContainer" containerID="13d27d812bf2e04ba7894fd3d40f834e035ea6cc1d6d6bd4adb237611ec7d7c4" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.582140 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gdwpn"] Feb 27 19:41:10 crc kubenswrapper[4839]: W0227 19:41:10.588543 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccee8c8e_aa9b_4aa4_8f0d_03e56a3c33c7.slice/crio-fec84bdb532d856c67e7dd6b6840bcfe2c4bbbb0bc195a23c7213afc1a2f579a WatchSource:0}: Error finding container fec84bdb532d856c67e7dd6b6840bcfe2c4bbbb0bc195a23c7213afc1a2f579a: Status 404 returned error can't find the container with id fec84bdb532d856c67e7dd6b6840bcfe2c4bbbb0bc195a23c7213afc1a2f579a Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.597140 4839 scope.go:117] "RemoveContainer" containerID="34b26e6b526bc8556137f5a0bcc2a8c24bc7556d1dc4bf20508e5d39ed9924d3" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.610635 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fddl"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.630646 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fddl"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.641686 4839 scope.go:117] "RemoveContainer" containerID="6f6bc15dd0710573d6c72e4b110fae722d4d86042550b9431787d9bc0a0e9d34" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.643378 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-clhjs"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.649478 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.649508 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a8a402-5d27-4632-8cf2-aaccccfff95d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.649524 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n76bg\" (UniqueName: \"kubernetes.io/projected/68a8a402-5d27-4632-8cf2-aaccccfff95d-kube-api-access-n76bg\") on node \"crc\" DevicePath \"\"" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.651030 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-clhjs"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.661535 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-stz5c"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.666239 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-stz5c"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.668887 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m6qmh"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.672266 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m6qmh"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.886473 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff888"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.891000 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ff888"] Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.971967 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" path="/var/lib/kubelet/pods/2cb7672b-b64f-4e33-9ea3-939a2a6e9da2/volumes" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.972433 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" path="/var/lib/kubelet/pods/67e667b9-304d-4052-8f0f-8d46d920aa7c/volumes" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.973016 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" path="/var/lib/kubelet/pods/68a8a402-5d27-4632-8cf2-aaccccfff95d/volumes" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.974028 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" path="/var/lib/kubelet/pods/b471c70a-15fe-4b5f-aa6f-354848925ace/volumes" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.974863 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f11b8271-1de0-432d-8d78-2379862726a1" path="/var/lib/kubelet/pods/f11b8271-1de0-432d-8d78-2379862726a1/volumes" Feb 27 19:41:10 crc kubenswrapper[4839]: I0227 19:41:10.975787 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" path="/var/lib/kubelet/pods/ff0a141c-3800-4c84-b1d6-17ec909d5d27/volumes" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267454 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mngj4"] Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267716 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267732 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267750 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267758 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267773 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerName="marketplace-operator" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267782 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerName="marketplace-operator" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267795 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267802 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267811 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267819 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267827 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267835 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267846 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267855 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267865 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267872 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267883 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267902 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267913 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267920 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267933 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267940 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267951 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267960 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.267970 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.267978 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: E0227 19:41:11.268001 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268010 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="extract-content" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268142 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a8a402-5d27-4632-8cf2-aaccccfff95d" containerName="extract-utilities" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268158 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb7672b-b64f-4e33-9ea3-939a2a6e9da2" containerName="marketplace-operator" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268170 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0a141c-3800-4c84-b1d6-17ec909d5d27" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268181 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e667b9-304d-4052-8f0f-8d46d920aa7c" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268191 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="f11b8271-1de0-432d-8d78-2379862726a1" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.268202 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="b471c70a-15fe-4b5f-aa6f-354848925ace" containerName="registry-server" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.269044 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.270879 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.276960 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mngj4"] Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.358290 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babba9f2-c281-4ad8-b697-9eb744a1960e-catalog-content\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.358353 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pfgp\" (UniqueName: \"kubernetes.io/projected/babba9f2-c281-4ad8-b697-9eb744a1960e-kube-api-access-9pfgp\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.358487 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babba9f2-c281-4ad8-b697-9eb744a1960e-utilities\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.459263 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babba9f2-c281-4ad8-b697-9eb744a1960e-utilities\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.459677 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babba9f2-c281-4ad8-b697-9eb744a1960e-catalog-content\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.459706 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pfgp\" (UniqueName: \"kubernetes.io/projected/babba9f2-c281-4ad8-b697-9eb744a1960e-kube-api-access-9pfgp\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.459969 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babba9f2-c281-4ad8-b697-9eb744a1960e-utilities\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.460107 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babba9f2-c281-4ad8-b697-9eb744a1960e-catalog-content\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.466226 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fqhwp"] Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.469999 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.472953 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.474267 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqhwp"] Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.490092 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pfgp\" (UniqueName: \"kubernetes.io/projected/babba9f2-c281-4ad8-b697-9eb744a1960e-kube-api-access-9pfgp\") pod \"redhat-marketplace-mngj4\" (UID: \"babba9f2-c281-4ad8-b697-9eb744a1960e\") " pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.548601 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" event={"ID":"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7","Type":"ContainerStarted","Data":"9477b3a8b2ced6acb2d7032bedfdad965e5329079ae7eb4ee7551ab5bbdb2f48"} Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.548646 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" event={"ID":"ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7","Type":"ContainerStarted","Data":"fec84bdb532d856c67e7dd6b6840bcfe2c4bbbb0bc195a23c7213afc1a2f579a"} Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.549925 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.554279 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.560420 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj5wr\" (UniqueName: \"kubernetes.io/projected/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-kube-api-access-jj5wr\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.560472 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-catalog-content\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.560491 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-utilities\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.586255 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gdwpn" podStartSLOduration=2.5862397660000003 podStartE2EDuration="2.586239766s" podCreationTimestamp="2026-02-27 19:41:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:41:11.567143907 +0000 UTC m=+433.212013652" watchObservedRunningTime="2026-02-27 19:41:11.586239766 +0000 UTC m=+433.231109501" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.594769 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mngj4" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.661258 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj5wr\" (UniqueName: \"kubernetes.io/projected/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-kube-api-access-jj5wr\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.661310 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-catalog-content\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.661331 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-utilities\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.661812 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-utilities\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.661821 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-catalog-content\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.679848 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj5wr\" (UniqueName: \"kubernetes.io/projected/e640d992-0a39-4fd3-b53b-5a8afe9a5a4a-kube-api-access-jj5wr\") pod \"certified-operators-fqhwp\" (UID: \"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a\") " pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.816342 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.979830 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mngj4"] Feb 27 19:41:11 crc kubenswrapper[4839]: I0227 19:41:11.983457 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqhwp"] Feb 27 19:41:11 crc kubenswrapper[4839]: W0227 19:41:11.984333 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbabba9f2_c281_4ad8_b697_9eb744a1960e.slice/crio-502419bc19f091b1b1b0f1940db2efc8ddcbcafd77e15d3bd575ead747de5832 WatchSource:0}: Error finding container 502419bc19f091b1b1b0f1940db2efc8ddcbcafd77e15d3bd575ead747de5832: Status 404 returned error can't find the container with id 502419bc19f091b1b1b0f1940db2efc8ddcbcafd77e15d3bd575ead747de5832 Feb 27 19:41:12 crc kubenswrapper[4839]: I0227 19:41:12.557966 4839 generic.go:334] "Generic (PLEG): container finished" podID="e640d992-0a39-4fd3-b53b-5a8afe9a5a4a" containerID="5a11bfa809e9fc31a40660913a9e97f5e5b94dce765038f4f0aee260864e5b53" exitCode=0 Feb 27 19:41:12 crc kubenswrapper[4839]: I0227 19:41:12.558070 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqhwp" event={"ID":"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a","Type":"ContainerDied","Data":"5a11bfa809e9fc31a40660913a9e97f5e5b94dce765038f4f0aee260864e5b53"} Feb 27 19:41:12 crc kubenswrapper[4839]: I0227 19:41:12.558141 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqhwp" event={"ID":"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a","Type":"ContainerStarted","Data":"11420c02b7f0fa6bc80559bae17f68ddf8bf54392ed161fc18ac38e66322aae5"} Feb 27 19:41:12 crc kubenswrapper[4839]: I0227 19:41:12.560705 4839 generic.go:334] "Generic (PLEG): container finished" podID="babba9f2-c281-4ad8-b697-9eb744a1960e" containerID="eb342e31d2b3c5e5580dbc74a1c42a3554696e407c466b7e3f1df11dd6ab7696" exitCode=0 Feb 27 19:41:12 crc kubenswrapper[4839]: I0227 19:41:12.560744 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mngj4" event={"ID":"babba9f2-c281-4ad8-b697-9eb744a1960e","Type":"ContainerDied","Data":"eb342e31d2b3c5e5580dbc74a1c42a3554696e407c466b7e3f1df11dd6ab7696"} Feb 27 19:41:12 crc kubenswrapper[4839]: I0227 19:41:12.560821 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mngj4" event={"ID":"babba9f2-c281-4ad8-b697-9eb744a1960e","Type":"ContainerStarted","Data":"502419bc19f091b1b1b0f1940db2efc8ddcbcafd77e15d3bd575ead747de5832"} Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.567365 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqhwp" event={"ID":"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a","Type":"ContainerStarted","Data":"6a19cc911cbc9f4f310000be2ab87583b6da1e95e7fe002e15e1c93fde7ccedf"} Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.662836 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f5dwg"] Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.663760 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.667124 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.675648 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5dwg"] Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.806383 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7986d862-dbf4-452a-a1e7-2d02cc850d76-catalog-content\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.806439 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7986d862-dbf4-452a-a1e7-2d02cc850d76-utilities\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.806499 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76zwd\" (UniqueName: \"kubernetes.io/projected/7986d862-dbf4-452a-a1e7-2d02cc850d76-kube-api-access-76zwd\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.866768 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7vmj"] Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.868162 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.869717 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.879108 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7vmj"] Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.908238 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76zwd\" (UniqueName: \"kubernetes.io/projected/7986d862-dbf4-452a-a1e7-2d02cc850d76-kube-api-access-76zwd\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.908335 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7986d862-dbf4-452a-a1e7-2d02cc850d76-catalog-content\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.908397 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7986d862-dbf4-452a-a1e7-2d02cc850d76-utilities\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.909143 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7986d862-dbf4-452a-a1e7-2d02cc850d76-catalog-content\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.909299 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7986d862-dbf4-452a-a1e7-2d02cc850d76-utilities\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:13 crc kubenswrapper[4839]: I0227 19:41:13.943697 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76zwd\" (UniqueName: \"kubernetes.io/projected/7986d862-dbf4-452a-a1e7-2d02cc850d76-kube-api-access-76zwd\") pod \"community-operators-f5dwg\" (UID: \"7986d862-dbf4-452a-a1e7-2d02cc850d76\") " pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.009197 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/827f2124-61ed-4c5e-acf7-21463af06877-utilities\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.009237 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/827f2124-61ed-4c5e-acf7-21463af06877-catalog-content\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.009357 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l8rn\" (UniqueName: \"kubernetes.io/projected/827f2124-61ed-4c5e-acf7-21463af06877-kube-api-access-8l8rn\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.023480 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.110309 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l8rn\" (UniqueName: \"kubernetes.io/projected/827f2124-61ed-4c5e-acf7-21463af06877-kube-api-access-8l8rn\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.110407 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/827f2124-61ed-4c5e-acf7-21463af06877-utilities\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.110433 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/827f2124-61ed-4c5e-acf7-21463af06877-catalog-content\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.110913 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/827f2124-61ed-4c5e-acf7-21463af06877-catalog-content\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.111170 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/827f2124-61ed-4c5e-acf7-21463af06877-utilities\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.131245 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l8rn\" (UniqueName: \"kubernetes.io/projected/827f2124-61ed-4c5e-acf7-21463af06877-kube-api-access-8l8rn\") pod \"redhat-operators-x7vmj\" (UID: \"827f2124-61ed-4c5e-acf7-21463af06877\") " pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.191895 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7vmj" Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.222533 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5dwg"] Feb 27 19:41:14 crc kubenswrapper[4839]: W0227 19:41:14.226900 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7986d862_dbf4_452a_a1e7_2d02cc850d76.slice/crio-2ddc9e81fbd5c5f993820798b1f98bc026e138f1e451c2d580d17fbcf92faadd WatchSource:0}: Error finding container 2ddc9e81fbd5c5f993820798b1f98bc026e138f1e451c2d580d17fbcf92faadd: Status 404 returned error can't find the container with id 2ddc9e81fbd5c5f993820798b1f98bc026e138f1e451c2d580d17fbcf92faadd Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.575007 4839 generic.go:334] "Generic (PLEG): container finished" podID="e640d992-0a39-4fd3-b53b-5a8afe9a5a4a" containerID="6a19cc911cbc9f4f310000be2ab87583b6da1e95e7fe002e15e1c93fde7ccedf" exitCode=0 Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.575104 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqhwp" event={"ID":"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a","Type":"ContainerDied","Data":"6a19cc911cbc9f4f310000be2ab87583b6da1e95e7fe002e15e1c93fde7ccedf"} Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.578872 4839 generic.go:334] "Generic (PLEG): container finished" podID="7986d862-dbf4-452a-a1e7-2d02cc850d76" containerID="58cfb26fe4269de929c9719969d512f5a9d0404710ce6259967099abe4d0bf0d" exitCode=0 Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.578915 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5dwg" event={"ID":"7986d862-dbf4-452a-a1e7-2d02cc850d76","Type":"ContainerDied","Data":"58cfb26fe4269de929c9719969d512f5a9d0404710ce6259967099abe4d0bf0d"} Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.578941 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5dwg" event={"ID":"7986d862-dbf4-452a-a1e7-2d02cc850d76","Type":"ContainerStarted","Data":"2ddc9e81fbd5c5f993820798b1f98bc026e138f1e451c2d580d17fbcf92faadd"} Feb 27 19:41:14 crc kubenswrapper[4839]: I0227 19:41:14.607384 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7vmj"] Feb 27 19:41:14 crc kubenswrapper[4839]: W0227 19:41:14.634125 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod827f2124_61ed_4c5e_acf7_21463af06877.slice/crio-07f56c85a9eb258c239a754858a855fb453d74262516d130c9faf1e14bac1e28 WatchSource:0}: Error finding container 07f56c85a9eb258c239a754858a855fb453d74262516d130c9faf1e14bac1e28: Status 404 returned error can't find the container with id 07f56c85a9eb258c239a754858a855fb453d74262516d130c9faf1e14bac1e28 Feb 27 19:41:15 crc kubenswrapper[4839]: E0227 19:41:15.307394 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:41:15 crc kubenswrapper[4839]: E0227 19:41:15.307877 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:41:15 crc kubenswrapper[4839]: E0227 19:41:15.309069 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:41:15 crc kubenswrapper[4839]: I0227 19:41:15.593758 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5dwg" event={"ID":"7986d862-dbf4-452a-a1e7-2d02cc850d76","Type":"ContainerStarted","Data":"a091fa5a32d09e857717918932de0aa5485433c90a028bd7be8fa2a7ab0fa9fe"} Feb 27 19:41:15 crc kubenswrapper[4839]: I0227 19:41:15.597237 4839 generic.go:334] "Generic (PLEG): container finished" podID="827f2124-61ed-4c5e-acf7-21463af06877" containerID="511aac28f285fc01d3e31e50b6383944763fec34dfcb8df20e92b8da6373742b" exitCode=0 Feb 27 19:41:15 crc kubenswrapper[4839]: I0227 19:41:15.597284 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vmj" event={"ID":"827f2124-61ed-4c5e-acf7-21463af06877","Type":"ContainerDied","Data":"511aac28f285fc01d3e31e50b6383944763fec34dfcb8df20e92b8da6373742b"} Feb 27 19:41:15 crc kubenswrapper[4839]: I0227 19:41:15.597302 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vmj" event={"ID":"827f2124-61ed-4c5e-acf7-21463af06877","Type":"ContainerStarted","Data":"07f56c85a9eb258c239a754858a855fb453d74262516d130c9faf1e14bac1e28"} Feb 27 19:41:15 crc kubenswrapper[4839]: I0227 19:41:15.604990 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqhwp" event={"ID":"e640d992-0a39-4fd3-b53b-5a8afe9a5a4a","Type":"ContainerStarted","Data":"ce6eb1b7a7fc2669d30c58ada9a21deda8b0566c0f4733284831369c9648954e"} Feb 27 19:41:15 crc kubenswrapper[4839]: E0227 19:41:15.606655 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:41:15 crc kubenswrapper[4839]: I0227 19:41:15.671779 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fqhwp" podStartSLOduration=1.882687797 podStartE2EDuration="4.67176449s" podCreationTimestamp="2026-02-27 19:41:11 +0000 UTC" firstStartedPulling="2026-02-27 19:41:12.560030712 +0000 UTC m=+434.204900447" lastFinishedPulling="2026-02-27 19:41:15.349107405 +0000 UTC m=+436.993977140" observedRunningTime="2026-02-27 19:41:15.666853372 +0000 UTC m=+437.311723127" watchObservedRunningTime="2026-02-27 19:41:15.67176449 +0000 UTC m=+437.316634225" Feb 27 19:41:16 crc kubenswrapper[4839]: E0227 19:41:16.242692 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:41:16 crc kubenswrapper[4839]: E0227 19:41:16.242874 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:41:16 crc kubenswrapper[4839]: E0227 19:41:16.244321 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:41:16 crc kubenswrapper[4839]: I0227 19:41:16.612585 4839 generic.go:334] "Generic (PLEG): container finished" podID="7986d862-dbf4-452a-a1e7-2d02cc850d76" containerID="a091fa5a32d09e857717918932de0aa5485433c90a028bd7be8fa2a7ab0fa9fe" exitCode=0 Feb 27 19:41:16 crc kubenswrapper[4839]: I0227 19:41:16.612619 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5dwg" event={"ID":"7986d862-dbf4-452a-a1e7-2d02cc850d76","Type":"ContainerDied","Data":"a091fa5a32d09e857717918932de0aa5485433c90a028bd7be8fa2a7ab0fa9fe"} Feb 27 19:41:16 crc kubenswrapper[4839]: E0227 19:41:16.614626 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:41:17 crc kubenswrapper[4839]: I0227 19:41:17.620432 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5dwg" event={"ID":"7986d862-dbf4-452a-a1e7-2d02cc850d76","Type":"ContainerStarted","Data":"169e59cff5f93184bcace8c50e680e0fadbcbc02d55671229bf6a21d615d5f30"} Feb 27 19:41:17 crc kubenswrapper[4839]: I0227 19:41:17.638358 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f5dwg" podStartSLOduration=2.192634715 podStartE2EDuration="4.63834335s" podCreationTimestamp="2026-02-27 19:41:13 +0000 UTC" firstStartedPulling="2026-02-27 19:41:14.580285177 +0000 UTC m=+436.225154912" lastFinishedPulling="2026-02-27 19:41:17.025993812 +0000 UTC m=+438.670863547" observedRunningTime="2026-02-27 19:41:17.636213165 +0000 UTC m=+439.281082910" watchObservedRunningTime="2026-02-27 19:41:17.63834335 +0000 UTC m=+439.283213075" Feb 27 19:41:21 crc kubenswrapper[4839]: I0227 19:41:21.816454 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:21 crc kubenswrapper[4839]: I0227 19:41:21.816747 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:21 crc kubenswrapper[4839]: I0227 19:41:21.889894 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:22 crc kubenswrapper[4839]: I0227 19:41:22.682293 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:41:22 crc kubenswrapper[4839]: I0227 19:41:22.682354 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:41:22 crc kubenswrapper[4839]: I0227 19:41:22.688437 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fqhwp" Feb 27 19:41:24 crc kubenswrapper[4839]: I0227 19:41:24.023917 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:24 crc kubenswrapper[4839]: I0227 19:41:24.024291 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:24 crc kubenswrapper[4839]: I0227 19:41:24.081300 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:24 crc kubenswrapper[4839]: I0227 19:41:24.702049 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f5dwg" Feb 27 19:41:28 crc kubenswrapper[4839]: E0227 19:41:28.804629 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:41:28 crc kubenswrapper[4839]: E0227 19:41:28.805852 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:41:28 crc kubenswrapper[4839]: E0227 19:41:28.807086 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:41:30 crc kubenswrapper[4839]: E0227 19:41:30.523806 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:41:30 crc kubenswrapper[4839]: E0227 19:41:30.523990 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:41:30 crc kubenswrapper[4839]: E0227 19:41:30.525234 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:41:39 crc kubenswrapper[4839]: E0227 19:41:39.968563 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:41:40 crc kubenswrapper[4839]: E0227 19:41:40.968842 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:41:52 crc kubenswrapper[4839]: E0227 19:41:52.634646 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:41:52 crc kubenswrapper[4839]: E0227 19:41:52.635931 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:41:52 crc kubenswrapper[4839]: E0227 19:41:52.637224 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.682011 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.682077 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.682127 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.682730 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d66dbb8b18e6a5a083cdeabb414a44ff334ee6690e03ac1dc87884b44ab7d3f2"} pod="openshift-machine-config-operator/machine-config-daemon-thb8n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.682795 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" containerID="cri-o://d66dbb8b18e6a5a083cdeabb414a44ff334ee6690e03ac1dc87884b44ab7d3f2" gracePeriod=600 Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.821103 4839 generic.go:334] "Generic (PLEG): container finished" podID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerID="d66dbb8b18e6a5a083cdeabb414a44ff334ee6690e03ac1dc87884b44ab7d3f2" exitCode=0 Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.821161 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerDied","Data":"d66dbb8b18e6a5a083cdeabb414a44ff334ee6690e03ac1dc87884b44ab7d3f2"} Feb 27 19:41:52 crc kubenswrapper[4839]: I0227 19:41:52.821208 4839 scope.go:117] "RemoveContainer" containerID="f7b768d8024d63f26535cd845992706b303742e2792d842d3317e7751312da1b" Feb 27 19:41:53 crc kubenswrapper[4839]: I0227 19:41:53.828501 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"1b68014096a9cd142c188dd5e27a95973d3809ea3da725eb9ba31acfc05d2d48"} Feb 27 19:41:54 crc kubenswrapper[4839]: E0227 19:41:54.605594 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:41:54 crc kubenswrapper[4839]: E0227 19:41:54.605787 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:41:54 crc kubenswrapper[4839]: E0227 19:41:54.607048 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.142145 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537022-qgfbs"] Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.143552 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.146488 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44lp8" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.146705 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.146828 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.162967 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537022-qgfbs"] Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.253581 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44hl9\" (UniqueName: \"kubernetes.io/projected/36d9c59c-867a-41d5-b435-7c0693c857cb-kube-api-access-44hl9\") pod \"auto-csr-approver-29537022-qgfbs\" (UID: \"36d9c59c-867a-41d5-b435-7c0693c857cb\") " pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.355029 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44hl9\" (UniqueName: \"kubernetes.io/projected/36d9c59c-867a-41d5-b435-7c0693c857cb-kube-api-access-44hl9\") pod \"auto-csr-approver-29537022-qgfbs\" (UID: \"36d9c59c-867a-41d5-b435-7c0693c857cb\") " pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.389338 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44hl9\" (UniqueName: \"kubernetes.io/projected/36d9c59c-867a-41d5-b435-7c0693c857cb-kube-api-access-44hl9\") pod \"auto-csr-approver-29537022-qgfbs\" (UID: \"36d9c59c-867a-41d5-b435-7c0693c857cb\") " pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.472998 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:42:00 crc kubenswrapper[4839]: I0227 19:42:00.907863 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537022-qgfbs"] Feb 27 19:42:00 crc kubenswrapper[4839]: W0227 19:42:00.919741 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36d9c59c_867a_41d5_b435_7c0693c857cb.slice/crio-e78759cba2a57d3e805a34ea04cd189c65f3f795a8806df1d6cddb97f04c0c65 WatchSource:0}: Error finding container e78759cba2a57d3e805a34ea04cd189c65f3f795a8806df1d6cddb97f04c0c65: Status 404 returned error can't find the container with id e78759cba2a57d3e805a34ea04cd189c65f3f795a8806df1d6cddb97f04c0c65 Feb 27 19:42:01 crc kubenswrapper[4839]: I0227 19:42:01.878964 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" event={"ID":"36d9c59c-867a-41d5-b435-7c0693c857cb","Type":"ContainerStarted","Data":"e78759cba2a57d3e805a34ea04cd189c65f3f795a8806df1d6cddb97f04c0c65"} Feb 27 19:42:02 crc kubenswrapper[4839]: E0227 19:42:02.455097 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:42:02 crc kubenswrapper[4839]: E0227 19:42:02.455254 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:42:02 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:42:02 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44hl9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537022-qgfbs_openshift-infra(36d9c59c-867a-41d5-b435-7c0693c857cb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:42:02 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:42:02 crc kubenswrapper[4839]: E0227 19:42:02.457217 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:42:02 crc kubenswrapper[4839]: E0227 19:42:02.887186 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:42:06 crc kubenswrapper[4839]: E0227 19:42:06.969189 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:42:08 crc kubenswrapper[4839]: E0227 19:42:08.971306 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:42:16 crc kubenswrapper[4839]: E0227 19:42:16.647993 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:42:16 crc kubenswrapper[4839]: E0227 19:42:16.648455 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:42:16 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:42:16 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44hl9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537022-qgfbs_openshift-infra(36d9c59c-867a-41d5-b435-7c0693c857cb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:42:16 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:42:16 crc kubenswrapper[4839]: E0227 19:42:16.649782 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:42:19 crc kubenswrapper[4839]: E0227 19:42:19.968778 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:42:20 crc kubenswrapper[4839]: E0227 19:42:20.967083 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:42:31 crc kubenswrapper[4839]: E0227 19:42:31.969225 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:42:31 crc kubenswrapper[4839]: E0227 19:42:31.969253 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:42:33 crc kubenswrapper[4839]: E0227 19:42:33.967204 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:42:45 crc kubenswrapper[4839]: I0227 19:42:45.969397 4839 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 19:42:47 crc kubenswrapper[4839]: E0227 19:42:47.491017 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:42:47 crc kubenswrapper[4839]: E0227 19:42:47.491572 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:42:47 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:42:47 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44hl9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537022-qgfbs_openshift-infra(36d9c59c-867a-41d5-b435-7c0693c857cb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:42:47 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:42:47 crc kubenswrapper[4839]: E0227 19:42:47.493556 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:42:48 crc kubenswrapper[4839]: E0227 19:42:48.058653 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:42:48 crc kubenswrapper[4839]: E0227 19:42:48.058890 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:42:48 crc kubenswrapper[4839]: E0227 19:42:48.060198 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:42:48 crc kubenswrapper[4839]: E0227 19:42:48.073583 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:42:48 crc kubenswrapper[4839]: E0227 19:42:48.073819 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:42:48 crc kubenswrapper[4839]: E0227 19:42:48.075057 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:42:58 crc kubenswrapper[4839]: E0227 19:42:58.975703 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:42:59 crc kubenswrapper[4839]: E0227 19:42:59.968307 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:43:01 crc kubenswrapper[4839]: E0227 19:43:01.967363 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:43:10 crc kubenswrapper[4839]: E0227 19:43:10.967853 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:43:14 crc kubenswrapper[4839]: E0227 19:43:14.969892 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:43:16 crc kubenswrapper[4839]: E0227 19:43:16.967834 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:43:24 crc kubenswrapper[4839]: E0227 19:43:24.969022 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:43:26 crc kubenswrapper[4839]: E0227 19:43:26.969441 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:43:31 crc kubenswrapper[4839]: E0227 19:43:31.967704 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:43:39 crc kubenswrapper[4839]: E0227 19:43:39.968365 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:43:41 crc kubenswrapper[4839]: E0227 19:43:41.765758 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:43:41 crc kubenswrapper[4839]: E0227 19:43:41.766373 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:43:41 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:43:41 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44hl9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537022-qgfbs_openshift-infra(36d9c59c-867a-41d5-b435-7c0693c857cb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:43:41 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:43:41 crc kubenswrapper[4839]: E0227 19:43:41.767623 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:43:46 crc kubenswrapper[4839]: E0227 19:43:46.967839 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:43:52 crc kubenswrapper[4839]: I0227 19:43:52.709098 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:43:52 crc kubenswrapper[4839]: I0227 19:43:52.709174 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:43:52 crc kubenswrapper[4839]: E0227 19:43:52.967592 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:43:53 crc kubenswrapper[4839]: E0227 19:43:53.967948 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:43:59 crc kubenswrapper[4839]: I0227 19:43:59.263485 4839 scope.go:117] "RemoveContainer" containerID="af4d438849daa5d4e132ae67e3659b425f69071f5b4262fbdd5a5f7837f9cc04" Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.153682 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537024-szhk2"] Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.156646 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537024-szhk2"] Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.156759 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.300434 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snvj8\" (UniqueName: \"kubernetes.io/projected/45766657-34fd-4b18-a32b-8e71a1c711e2-kube-api-access-snvj8\") pod \"auto-csr-approver-29537024-szhk2\" (UID: \"45766657-34fd-4b18-a32b-8e71a1c711e2\") " pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.402256 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snvj8\" (UniqueName: \"kubernetes.io/projected/45766657-34fd-4b18-a32b-8e71a1c711e2-kube-api-access-snvj8\") pod \"auto-csr-approver-29537024-szhk2\" (UID: \"45766657-34fd-4b18-a32b-8e71a1c711e2\") " pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.429008 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snvj8\" (UniqueName: \"kubernetes.io/projected/45766657-34fd-4b18-a32b-8e71a1c711e2-kube-api-access-snvj8\") pod \"auto-csr-approver-29537024-szhk2\" (UID: \"45766657-34fd-4b18-a32b-8e71a1c711e2\") " pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.514766 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:00 crc kubenswrapper[4839]: I0227 19:44:00.745820 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537024-szhk2"] Feb 27 19:44:00 crc kubenswrapper[4839]: W0227 19:44:00.750143 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45766657_34fd_4b18_a32b_8e71a1c711e2.slice/crio-86fcdba530baee7eb5b8e5bd08842baa63bf01e86a793e94a82907d8f0b4b3d5 WatchSource:0}: Error finding container 86fcdba530baee7eb5b8e5bd08842baa63bf01e86a793e94a82907d8f0b4b3d5: Status 404 returned error can't find the container with id 86fcdba530baee7eb5b8e5bd08842baa63bf01e86a793e94a82907d8f0b4b3d5 Feb 27 19:44:00 crc kubenswrapper[4839]: E0227 19:44:00.968170 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:44:01 crc kubenswrapper[4839]: I0227 19:44:01.619962 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537024-szhk2" event={"ID":"45766657-34fd-4b18-a32b-8e71a1c711e2","Type":"ContainerStarted","Data":"86fcdba530baee7eb5b8e5bd08842baa63bf01e86a793e94a82907d8f0b4b3d5"} Feb 27 19:44:01 crc kubenswrapper[4839]: E0227 19:44:01.638321 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:44:01 crc kubenswrapper[4839]: E0227 19:44:01.638550 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:44:01 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:44:01 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-snvj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537024-szhk2_openshift-infra(45766657-34fd-4b18-a32b-8e71a1c711e2): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:44:01 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:44:01 crc kubenswrapper[4839]: E0227 19:44:01.640001 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537024-szhk2" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" Feb 27 19:44:02 crc kubenswrapper[4839]: E0227 19:44:02.627415 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537024-szhk2" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" Feb 27 19:44:03 crc kubenswrapper[4839]: E0227 19:44:03.968075 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:44:05 crc kubenswrapper[4839]: E0227 19:44:05.980719 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:44:14 crc kubenswrapper[4839]: E0227 19:44:14.968974 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:44:16 crc kubenswrapper[4839]: E0227 19:44:16.670755 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:44:16 crc kubenswrapper[4839]: E0227 19:44:16.671191 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:44:16 crc kubenswrapper[4839]: E0227 19:44:16.672967 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:44:16 crc kubenswrapper[4839]: E0227 19:44:16.923453 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:44:16 crc kubenswrapper[4839]: E0227 19:44:16.923926 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:44:16 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:44:16 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-snvj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537024-szhk2_openshift-infra(45766657-34fd-4b18-a32b-8e71a1c711e2): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:44:16 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:44:16 crc kubenswrapper[4839]: E0227 19:44:16.925345 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537024-szhk2" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" Feb 27 19:44:19 crc kubenswrapper[4839]: E0227 19:44:19.561110 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:44:19 crc kubenswrapper[4839]: E0227 19:44:19.561853 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:44:19 crc kubenswrapper[4839]: E0227 19:44:19.563398 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:44:22 crc kubenswrapper[4839]: I0227 19:44:22.682526 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:44:22 crc kubenswrapper[4839]: I0227 19:44:22.683899 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:44:26 crc kubenswrapper[4839]: E0227 19:44:26.969843 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:44:27 crc kubenswrapper[4839]: E0227 19:44:27.966902 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:44:30 crc kubenswrapper[4839]: E0227 19:44:30.967980 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537024-szhk2" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" Feb 27 19:44:31 crc kubenswrapper[4839]: E0227 19:44:31.967083 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:44:40 crc kubenswrapper[4839]: E0227 19:44:40.968506 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:44:40 crc kubenswrapper[4839]: E0227 19:44:40.969026 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:44:43 crc kubenswrapper[4839]: E0227 19:44:43.967782 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:44:47 crc kubenswrapper[4839]: I0227 19:44:47.920081 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537024-szhk2" event={"ID":"45766657-34fd-4b18-a32b-8e71a1c711e2","Type":"ContainerStarted","Data":"fd4d7d181beb069c7d7c147ddeb49b7444289e329deb312d5a53757520498513"} Feb 27 19:44:47 crc kubenswrapper[4839]: I0227 19:44:47.945255 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537024-szhk2" podStartSLOduration=1.1186883 podStartE2EDuration="47.945223413s" podCreationTimestamp="2026-02-27 19:44:00 +0000 UTC" firstStartedPulling="2026-02-27 19:44:00.753336981 +0000 UTC m=+602.398206706" lastFinishedPulling="2026-02-27 19:44:47.579872044 +0000 UTC m=+649.224741819" observedRunningTime="2026-02-27 19:44:47.939224674 +0000 UTC m=+649.584094439" watchObservedRunningTime="2026-02-27 19:44:47.945223413 +0000 UTC m=+649.590093178" Feb 27 19:44:48 crc kubenswrapper[4839]: I0227 19:44:48.928705 4839 generic.go:334] "Generic (PLEG): container finished" podID="45766657-34fd-4b18-a32b-8e71a1c711e2" containerID="fd4d7d181beb069c7d7c147ddeb49b7444289e329deb312d5a53757520498513" exitCode=0 Feb 27 19:44:48 crc kubenswrapper[4839]: I0227 19:44:48.928799 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537024-szhk2" event={"ID":"45766657-34fd-4b18-a32b-8e71a1c711e2","Type":"ContainerDied","Data":"fd4d7d181beb069c7d7c147ddeb49b7444289e329deb312d5a53757520498513"} Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.156148 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.287558 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snvj8\" (UniqueName: \"kubernetes.io/projected/45766657-34fd-4b18-a32b-8e71a1c711e2-kube-api-access-snvj8\") pod \"45766657-34fd-4b18-a32b-8e71a1c711e2\" (UID: \"45766657-34fd-4b18-a32b-8e71a1c711e2\") " Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.295222 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45766657-34fd-4b18-a32b-8e71a1c711e2-kube-api-access-snvj8" (OuterVolumeSpecName: "kube-api-access-snvj8") pod "45766657-34fd-4b18-a32b-8e71a1c711e2" (UID: "45766657-34fd-4b18-a32b-8e71a1c711e2"). InnerVolumeSpecName "kube-api-access-snvj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.389334 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snvj8\" (UniqueName: \"kubernetes.io/projected/45766657-34fd-4b18-a32b-8e71a1c711e2-kube-api-access-snvj8\") on node \"crc\" DevicePath \"\"" Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.946147 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537024-szhk2" event={"ID":"45766657-34fd-4b18-a32b-8e71a1c711e2","Type":"ContainerDied","Data":"86fcdba530baee7eb5b8e5bd08842baa63bf01e86a793e94a82907d8f0b4b3d5"} Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.946194 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86fcdba530baee7eb5b8e5bd08842baa63bf01e86a793e94a82907d8f0b4b3d5" Feb 27 19:44:50 crc kubenswrapper[4839]: I0227 19:44:50.946262 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537024-szhk2" Feb 27 19:44:51 crc kubenswrapper[4839]: I0227 19:44:51.012061 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537016-rt648"] Feb 27 19:44:51 crc kubenswrapper[4839]: I0227 19:44:51.018844 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537016-rt648"] Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.682016 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.682386 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.682439 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.683065 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b68014096a9cd142c188dd5e27a95973d3809ea3da725eb9ba31acfc05d2d48"} pod="openshift-machine-config-operator/machine-config-daemon-thb8n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.683127 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" containerID="cri-o://1b68014096a9cd142c188dd5e27a95973d3809ea3da725eb9ba31acfc05d2d48" gracePeriod=600 Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.961359 4839 generic.go:334] "Generic (PLEG): container finished" podID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerID="1b68014096a9cd142c188dd5e27a95973d3809ea3da725eb9ba31acfc05d2d48" exitCode=0 Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.961508 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerDied","Data":"1b68014096a9cd142c188dd5e27a95973d3809ea3da725eb9ba31acfc05d2d48"} Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.962006 4839 scope.go:117] "RemoveContainer" containerID="d66dbb8b18e6a5a083cdeabb414a44ff334ee6690e03ac1dc87884b44ab7d3f2" Feb 27 19:44:52 crc kubenswrapper[4839]: I0227 19:44:52.981104 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e89fb2-3271-46bd-878b-6ff642b17214" path="/var/lib/kubelet/pods/95e89fb2-3271-46bd-878b-6ff642b17214/volumes" Feb 27 19:44:53 crc kubenswrapper[4839]: E0227 19:44:53.967922 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:44:53 crc kubenswrapper[4839]: I0227 19:44:53.973696 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"5fc60b748d670020094881ec089d53e0f67db6413e9b475783fdc7432a738c18"} Feb 27 19:44:55 crc kubenswrapper[4839]: E0227 19:44:55.967548 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" Feb 27 19:44:57 crc kubenswrapper[4839]: E0227 19:44:57.968284 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.156321 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49"] Feb 27 19:45:00 crc kubenswrapper[4839]: E0227 19:45:00.156741 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" containerName="oc" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.156763 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" containerName="oc" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.156941 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" containerName="oc" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.157583 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.161395 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.161449 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.169989 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49"] Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.313595 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b83e08-a45d-4b2a-b544-04251ef384ec-config-volume\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.313731 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b83e08-a45d-4b2a-b544-04251ef384ec-secret-volume\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.313766 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhb5p\" (UniqueName: \"kubernetes.io/projected/06b83e08-a45d-4b2a-b544-04251ef384ec-kube-api-access-zhb5p\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.415063 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b83e08-a45d-4b2a-b544-04251ef384ec-secret-volume\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.415133 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhb5p\" (UniqueName: \"kubernetes.io/projected/06b83e08-a45d-4b2a-b544-04251ef384ec-kube-api-access-zhb5p\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.415271 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b83e08-a45d-4b2a-b544-04251ef384ec-config-volume\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.416677 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b83e08-a45d-4b2a-b544-04251ef384ec-config-volume\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.424206 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b83e08-a45d-4b2a-b544-04251ef384ec-secret-volume\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.439729 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhb5p\" (UniqueName: \"kubernetes.io/projected/06b83e08-a45d-4b2a-b544-04251ef384ec-kube-api-access-zhb5p\") pod \"collect-profiles-29537025-6sx49\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.487720 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:00 crc kubenswrapper[4839]: I0227 19:45:00.724369 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49"] Feb 27 19:45:00 crc kubenswrapper[4839]: W0227 19:45:00.728083 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06b83e08_a45d_4b2a_b544_04251ef384ec.slice/crio-d0ad3ce1bec8c51ee72c7e057590d0ef47dfc183bc7b6601a275e4955095dcf9 WatchSource:0}: Error finding container d0ad3ce1bec8c51ee72c7e057590d0ef47dfc183bc7b6601a275e4955095dcf9: Status 404 returned error can't find the container with id d0ad3ce1bec8c51ee72c7e057590d0ef47dfc183bc7b6601a275e4955095dcf9 Feb 27 19:45:01 crc kubenswrapper[4839]: I0227 19:45:01.017131 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" event={"ID":"06b83e08-a45d-4b2a-b544-04251ef384ec","Type":"ContainerStarted","Data":"d46caa48ab376935885af8ce99b16d5d823a4d3fdde1dd9431f06bc1c140c6fe"} Feb 27 19:45:01 crc kubenswrapper[4839]: I0227 19:45:01.017528 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" event={"ID":"06b83e08-a45d-4b2a-b544-04251ef384ec","Type":"ContainerStarted","Data":"d0ad3ce1bec8c51ee72c7e057590d0ef47dfc183bc7b6601a275e4955095dcf9"} Feb 27 19:45:02 crc kubenswrapper[4839]: I0227 19:45:02.026529 4839 generic.go:334] "Generic (PLEG): container finished" podID="06b83e08-a45d-4b2a-b544-04251ef384ec" containerID="d46caa48ab376935885af8ce99b16d5d823a4d3fdde1dd9431f06bc1c140c6fe" exitCode=0 Feb 27 19:45:02 crc kubenswrapper[4839]: I0227 19:45:02.026601 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" event={"ID":"06b83e08-a45d-4b2a-b544-04251ef384ec","Type":"ContainerDied","Data":"d46caa48ab376935885af8ce99b16d5d823a4d3fdde1dd9431f06bc1c140c6fe"} Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.323181 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.454290 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b83e08-a45d-4b2a-b544-04251ef384ec-config-volume\") pod \"06b83e08-a45d-4b2a-b544-04251ef384ec\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.454403 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b83e08-a45d-4b2a-b544-04251ef384ec-secret-volume\") pod \"06b83e08-a45d-4b2a-b544-04251ef384ec\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.454456 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhb5p\" (UniqueName: \"kubernetes.io/projected/06b83e08-a45d-4b2a-b544-04251ef384ec-kube-api-access-zhb5p\") pod \"06b83e08-a45d-4b2a-b544-04251ef384ec\" (UID: \"06b83e08-a45d-4b2a-b544-04251ef384ec\") " Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.455891 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b83e08-a45d-4b2a-b544-04251ef384ec-config-volume" (OuterVolumeSpecName: "config-volume") pod "06b83e08-a45d-4b2a-b544-04251ef384ec" (UID: "06b83e08-a45d-4b2a-b544-04251ef384ec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.460007 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06b83e08-a45d-4b2a-b544-04251ef384ec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "06b83e08-a45d-4b2a-b544-04251ef384ec" (UID: "06b83e08-a45d-4b2a-b544-04251ef384ec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.460796 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b83e08-a45d-4b2a-b544-04251ef384ec-kube-api-access-zhb5p" (OuterVolumeSpecName: "kube-api-access-zhb5p") pod "06b83e08-a45d-4b2a-b544-04251ef384ec" (UID: "06b83e08-a45d-4b2a-b544-04251ef384ec"). InnerVolumeSpecName "kube-api-access-zhb5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.556185 4839 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b83e08-a45d-4b2a-b544-04251ef384ec-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.556284 4839 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b83e08-a45d-4b2a-b544-04251ef384ec-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 19:45:03 crc kubenswrapper[4839]: I0227 19:45:03.556308 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhb5p\" (UniqueName: \"kubernetes.io/projected/06b83e08-a45d-4b2a-b544-04251ef384ec-kube-api-access-zhb5p\") on node \"crc\" DevicePath \"\"" Feb 27 19:45:04 crc kubenswrapper[4839]: I0227 19:45:04.046102 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" event={"ID":"06b83e08-a45d-4b2a-b544-04251ef384ec","Type":"ContainerDied","Data":"d0ad3ce1bec8c51ee72c7e057590d0ef47dfc183bc7b6601a275e4955095dcf9"} Feb 27 19:45:04 crc kubenswrapper[4839]: I0227 19:45:04.046171 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0ad3ce1bec8c51ee72c7e057590d0ef47dfc183bc7b6601a275e4955095dcf9" Feb 27 19:45:04 crc kubenswrapper[4839]: I0227 19:45:04.046262 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537025-6sx49" Feb 27 19:45:07 crc kubenswrapper[4839]: E0227 19:45:07.968637 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:45:10 crc kubenswrapper[4839]: I0227 19:45:10.079193 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" event={"ID":"36d9c59c-867a-41d5-b435-7c0693c857cb","Type":"ContainerStarted","Data":"bb374ab362b865917efe6077abab0be2ffca71aac8403dd4e8db06cda9145b4a"} Feb 27 19:45:10 crc kubenswrapper[4839]: I0227 19:45:10.093337 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" podStartSLOduration=1.170051673 podStartE2EDuration="3m10.093322392s" podCreationTimestamp="2026-02-27 19:42:00 +0000 UTC" firstStartedPulling="2026-02-27 19:42:00.923192551 +0000 UTC m=+482.568062296" lastFinishedPulling="2026-02-27 19:45:09.84646327 +0000 UTC m=+671.491333015" observedRunningTime="2026-02-27 19:45:10.092419698 +0000 UTC m=+671.737289433" watchObservedRunningTime="2026-02-27 19:45:10.093322392 +0000 UTC m=+671.738192127" Feb 27 19:45:11 crc kubenswrapper[4839]: I0227 19:45:11.088690 4839 generic.go:334] "Generic (PLEG): container finished" podID="36d9c59c-867a-41d5-b435-7c0693c857cb" containerID="bb374ab362b865917efe6077abab0be2ffca71aac8403dd4e8db06cda9145b4a" exitCode=0 Feb 27 19:45:11 crc kubenswrapper[4839]: I0227 19:45:11.088754 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" event={"ID":"36d9c59c-867a-41d5-b435-7c0693c857cb","Type":"ContainerDied","Data":"bb374ab362b865917efe6077abab0be2ffca71aac8403dd4e8db06cda9145b4a"} Feb 27 19:45:12 crc kubenswrapper[4839]: I0227 19:45:12.352859 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:45:12 crc kubenswrapper[4839]: I0227 19:45:12.513454 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44hl9\" (UniqueName: \"kubernetes.io/projected/36d9c59c-867a-41d5-b435-7c0693c857cb-kube-api-access-44hl9\") pod \"36d9c59c-867a-41d5-b435-7c0693c857cb\" (UID: \"36d9c59c-867a-41d5-b435-7c0693c857cb\") " Feb 27 19:45:12 crc kubenswrapper[4839]: I0227 19:45:12.519473 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d9c59c-867a-41d5-b435-7c0693c857cb-kube-api-access-44hl9" (OuterVolumeSpecName: "kube-api-access-44hl9") pod "36d9c59c-867a-41d5-b435-7c0693c857cb" (UID: "36d9c59c-867a-41d5-b435-7c0693c857cb"). InnerVolumeSpecName "kube-api-access-44hl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:45:12 crc kubenswrapper[4839]: I0227 19:45:12.615403 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44hl9\" (UniqueName: \"kubernetes.io/projected/36d9c59c-867a-41d5-b435-7c0693c857cb-kube-api-access-44hl9\") on node \"crc\" DevicePath \"\"" Feb 27 19:45:12 crc kubenswrapper[4839]: E0227 19:45:12.967715 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:45:13 crc kubenswrapper[4839]: I0227 19:45:13.104348 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" event={"ID":"36d9c59c-867a-41d5-b435-7c0693c857cb","Type":"ContainerDied","Data":"e78759cba2a57d3e805a34ea04cd189c65f3f795a8806df1d6cddb97f04c0c65"} Feb 27 19:45:13 crc kubenswrapper[4839]: I0227 19:45:13.104392 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e78759cba2a57d3e805a34ea04cd189c65f3f795a8806df1d6cddb97f04c0c65" Feb 27 19:45:13 crc kubenswrapper[4839]: I0227 19:45:13.104479 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537022-qgfbs" Feb 27 19:45:13 crc kubenswrapper[4839]: I0227 19:45:13.159916 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537018-dbqzs"] Feb 27 19:45:13 crc kubenswrapper[4839]: I0227 19:45:13.177603 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537018-dbqzs"] Feb 27 19:45:14 crc kubenswrapper[4839]: I0227 19:45:14.973576 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="499a12e9-ae64-4d99-93ed-e6af775581c0" path="/var/lib/kubelet/pods/499a12e9-ae64-4d99-93ed-e6af775581c0/volumes" Feb 27 19:45:18 crc kubenswrapper[4839]: E0227 19:45:18.969509 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:45:25 crc kubenswrapper[4839]: E0227 19:45:25.969946 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:45:31 crc kubenswrapper[4839]: E0227 19:45:31.967998 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:45:38 crc kubenswrapper[4839]: E0227 19:45:38.973314 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:45:45 crc kubenswrapper[4839]: E0227 19:45:45.967511 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:45:53 crc kubenswrapper[4839]: E0227 19:45:53.968105 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:45:58 crc kubenswrapper[4839]: E0227 19:45:58.971124 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:45:59 crc kubenswrapper[4839]: I0227 19:45:59.316319 4839 scope.go:117] "RemoveContainer" containerID="2f9e7c5935162e0ff84fc261b0127931500b22f6d1e95825491f070d9d629c4b" Feb 27 19:45:59 crc kubenswrapper[4839]: I0227 19:45:59.350498 4839 scope.go:117] "RemoveContainer" containerID="4091d93ec5f4af7327331baae1e03bbc9c9124f98cf2ba8eb41266edf46e5f29" Feb 27 19:45:59 crc kubenswrapper[4839]: I0227 19:45:59.368813 4839 scope.go:117] "RemoveContainer" containerID="a7dbef3da3ac32c0108174826e5f63ebb0f520792390c542863b864ef66fea9c" Feb 27 19:45:59 crc kubenswrapper[4839]: I0227 19:45:59.396233 4839 scope.go:117] "RemoveContainer" containerID="c062e48e1dcfbe0524b8bf04be3045ea174c4a5b81b1200ae146399c4db1bc29" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.152875 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537026-67hjl"] Feb 27 19:46:00 crc kubenswrapper[4839]: E0227 19:46:00.153793 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b83e08-a45d-4b2a-b544-04251ef384ec" containerName="collect-profiles" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.153827 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b83e08-a45d-4b2a-b544-04251ef384ec" containerName="collect-profiles" Feb 27 19:46:00 crc kubenswrapper[4839]: E0227 19:46:00.153875 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" containerName="oc" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.153913 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" containerName="oc" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.154186 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" containerName="oc" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.154239 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b83e08-a45d-4b2a-b544-04251ef384ec" containerName="collect-profiles" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.154953 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.158304 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.158490 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.158783 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44lp8" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.159567 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537026-67hjl"] Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.270393 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m46pl\" (UniqueName: \"kubernetes.io/projected/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b-kube-api-access-m46pl\") pod \"auto-csr-approver-29537026-67hjl\" (UID: \"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b\") " pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.371424 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m46pl\" (UniqueName: \"kubernetes.io/projected/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b-kube-api-access-m46pl\") pod \"auto-csr-approver-29537026-67hjl\" (UID: \"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b\") " pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.407085 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m46pl\" (UniqueName: \"kubernetes.io/projected/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b-kube-api-access-m46pl\") pod \"auto-csr-approver-29537026-67hjl\" (UID: \"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b\") " pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.477232 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:46:00 crc kubenswrapper[4839]: I0227 19:46:00.732879 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537026-67hjl"] Feb 27 19:46:00 crc kubenswrapper[4839]: W0227 19:46:00.740943 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5f3e68_002d_4f8f_a93a_b4854e58fc5b.slice/crio-f9c3d38bdfbb7425d422f105b01a646cd638d71310d6bc0d9662782093c13d58 WatchSource:0}: Error finding container f9c3d38bdfbb7425d422f105b01a646cd638d71310d6bc0d9662782093c13d58: Status 404 returned error can't find the container with id f9c3d38bdfbb7425d422f105b01a646cd638d71310d6bc0d9662782093c13d58 Feb 27 19:46:01 crc kubenswrapper[4839]: I0227 19:46:01.438811 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537026-67hjl" event={"ID":"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b","Type":"ContainerStarted","Data":"f9c3d38bdfbb7425d422f105b01a646cd638d71310d6bc0d9662782093c13d58"} Feb 27 19:46:01 crc kubenswrapper[4839]: E0227 19:46:01.694909 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:46:01 crc kubenswrapper[4839]: E0227 19:46:01.695098 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:46:01 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:46:01 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m46pl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537026-67hjl_openshift-infra(9b5f3e68-002d-4f8f-a93a-b4854e58fc5b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:46:01 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:46:01 crc kubenswrapper[4839]: E0227 19:46:01.697118 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:46:02 crc kubenswrapper[4839]: E0227 19:46:02.448249 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:46:07 crc kubenswrapper[4839]: E0227 19:46:07.970240 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:46:11 crc kubenswrapper[4839]: E0227 19:46:11.967528 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:46:16 crc kubenswrapper[4839]: E0227 19:46:16.050257 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:46:16 crc kubenswrapper[4839]: E0227 19:46:16.050817 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:46:16 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:46:16 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m46pl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537026-67hjl_openshift-infra(9b5f3e68-002d-4f8f-a93a-b4854e58fc5b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:46:16 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:46:16 crc kubenswrapper[4839]: E0227 19:46:16.052120 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:46:19 crc kubenswrapper[4839]: E0227 19:46:19.967929 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:46:24 crc kubenswrapper[4839]: E0227 19:46:24.969477 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:46:27 crc kubenswrapper[4839]: E0227 19:46:27.968169 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.824761 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-7rchl"] Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.826302 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.830765 4839 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-6h9kg" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.831007 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.831342 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-pd6vc"] Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.832289 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.832481 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-pd6vc" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.836118 4839 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-fct2l" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.842209 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-gxszm"] Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.843259 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.845479 4839 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-stbzf" Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.855092 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-7rchl"] Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.869825 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-pd6vc"] Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.873923 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-gxszm"] Feb 27 19:46:31 crc kubenswrapper[4839]: I0227 19:46:31.919905 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf2kw\" (UniqueName: \"kubernetes.io/projected/58e0aaec-ea1f-46f9-a599-7ead3e24341d-kube-api-access-wf2kw\") pod \"cert-manager-cainjector-cf98fcc89-7rchl\" (UID: \"58e0aaec-ea1f-46f9-a599-7ead3e24341d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.021180 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcvtw\" (UniqueName: \"kubernetes.io/projected/9c95be49-7215-4016-8d89-0dd008c7ba2f-kube-api-access-hcvtw\") pod \"cert-manager-858654f9db-pd6vc\" (UID: \"9c95be49-7215-4016-8d89-0dd008c7ba2f\") " pod="cert-manager/cert-manager-858654f9db-pd6vc" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.021256 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf2kw\" (UniqueName: \"kubernetes.io/projected/58e0aaec-ea1f-46f9-a599-7ead3e24341d-kube-api-access-wf2kw\") pod \"cert-manager-cainjector-cf98fcc89-7rchl\" (UID: \"58e0aaec-ea1f-46f9-a599-7ead3e24341d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.021279 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64mgq\" (UniqueName: \"kubernetes.io/projected/56781d07-f0fd-4d2e-97d4-784c15af5737-kube-api-access-64mgq\") pod \"cert-manager-webhook-687f57d79b-gxszm\" (UID: \"56781d07-f0fd-4d2e-97d4-784c15af5737\") " pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.044886 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf2kw\" (UniqueName: \"kubernetes.io/projected/58e0aaec-ea1f-46f9-a599-7ead3e24341d-kube-api-access-wf2kw\") pod \"cert-manager-cainjector-cf98fcc89-7rchl\" (UID: \"58e0aaec-ea1f-46f9-a599-7ead3e24341d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.122895 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcvtw\" (UniqueName: \"kubernetes.io/projected/9c95be49-7215-4016-8d89-0dd008c7ba2f-kube-api-access-hcvtw\") pod \"cert-manager-858654f9db-pd6vc\" (UID: \"9c95be49-7215-4016-8d89-0dd008c7ba2f\") " pod="cert-manager/cert-manager-858654f9db-pd6vc" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.123031 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64mgq\" (UniqueName: \"kubernetes.io/projected/56781d07-f0fd-4d2e-97d4-784c15af5737-kube-api-access-64mgq\") pod \"cert-manager-webhook-687f57d79b-gxszm\" (UID: \"56781d07-f0fd-4d2e-97d4-784c15af5737\") " pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.141820 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcvtw\" (UniqueName: \"kubernetes.io/projected/9c95be49-7215-4016-8d89-0dd008c7ba2f-kube-api-access-hcvtw\") pod \"cert-manager-858654f9db-pd6vc\" (UID: \"9c95be49-7215-4016-8d89-0dd008c7ba2f\") " pod="cert-manager/cert-manager-858654f9db-pd6vc" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.142753 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64mgq\" (UniqueName: \"kubernetes.io/projected/56781d07-f0fd-4d2e-97d4-784c15af5737-kube-api-access-64mgq\") pod \"cert-manager-webhook-687f57d79b-gxszm\" (UID: \"56781d07-f0fd-4d2e-97d4-784c15af5737\") " pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.164857 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.180247 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-pd6vc" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.190572 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.399407 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-gxszm"] Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.439460 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-pd6vc"] Feb 27 19:46:32 crc kubenswrapper[4839]: W0227 19:46:32.444763 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c95be49_7215_4016_8d89_0dd008c7ba2f.slice/crio-61a77484ba4178d602aa188a60d7a66657f8b2e5446ec5fdcfcec9f243adfc60 WatchSource:0}: Error finding container 61a77484ba4178d602aa188a60d7a66657f8b2e5446ec5fdcfcec9f243adfc60: Status 404 returned error can't find the container with id 61a77484ba4178d602aa188a60d7a66657f8b2e5446ec5fdcfcec9f243adfc60 Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.577460 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-7rchl"] Feb 27 19:46:32 crc kubenswrapper[4839]: W0227 19:46:32.582160 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58e0aaec_ea1f_46f9_a599_7ead3e24341d.slice/crio-4e71042097f6f89c1cd5be3b0f2d161cb0278318b17058203c62fdbdb840a19c WatchSource:0}: Error finding container 4e71042097f6f89c1cd5be3b0f2d161cb0278318b17058203c62fdbdb840a19c: Status 404 returned error can't find the container with id 4e71042097f6f89c1cd5be3b0f2d161cb0278318b17058203c62fdbdb840a19c Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.656134 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" event={"ID":"56781d07-f0fd-4d2e-97d4-784c15af5737","Type":"ContainerStarted","Data":"81242332a765e5a2e8a22c0743ca943753a85a43f2ef994fae737c5c1f2fd4f8"} Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.657312 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" event={"ID":"58e0aaec-ea1f-46f9-a599-7ead3e24341d","Type":"ContainerStarted","Data":"4e71042097f6f89c1cd5be3b0f2d161cb0278318b17058203c62fdbdb840a19c"} Feb 27 19:46:32 crc kubenswrapper[4839]: I0227 19:46:32.658264 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-pd6vc" event={"ID":"9c95be49-7215-4016-8d89-0dd008c7ba2f","Type":"ContainerStarted","Data":"61a77484ba4178d602aa188a60d7a66657f8b2e5446ec5fdcfcec9f243adfc60"} Feb 27 19:46:34 crc kubenswrapper[4839]: E0227 19:46:34.523411 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:46:35 crc kubenswrapper[4839]: I0227 19:46:35.673583 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-pd6vc" event={"ID":"9c95be49-7215-4016-8d89-0dd008c7ba2f","Type":"ContainerStarted","Data":"a1c581f48854f71c5c67f2aa88f108f2a502cd814ce611a0723e39d821174677"} Feb 27 19:46:35 crc kubenswrapper[4839]: I0227 19:46:35.688268 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-pd6vc" podStartSLOduration=2.429094808 podStartE2EDuration="4.688252156s" podCreationTimestamp="2026-02-27 19:46:31 +0000 UTC" firstStartedPulling="2026-02-27 19:46:32.446674625 +0000 UTC m=+754.091544360" lastFinishedPulling="2026-02-27 19:46:34.705831973 +0000 UTC m=+756.350701708" observedRunningTime="2026-02-27 19:46:35.686643673 +0000 UTC m=+757.331513428" watchObservedRunningTime="2026-02-27 19:46:35.688252156 +0000 UTC m=+757.333121891" Feb 27 19:46:36 crc kubenswrapper[4839]: I0227 19:46:36.682399 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" event={"ID":"56781d07-f0fd-4d2e-97d4-784c15af5737","Type":"ContainerStarted","Data":"9993edb8988c77a5661b3703ec1215f925db816bd22270abe6e3e028d03bdc89"} Feb 27 19:46:36 crc kubenswrapper[4839]: I0227 19:46:36.682946 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:36 crc kubenswrapper[4839]: I0227 19:46:36.684736 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" event={"ID":"58e0aaec-ea1f-46f9-a599-7ead3e24341d","Type":"ContainerStarted","Data":"cf3da178d9cbd8188fc358701d3b848e64882c6f777f51ac451fe6689d70d45b"} Feb 27 19:46:36 crc kubenswrapper[4839]: I0227 19:46:36.728152 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-7rchl" podStartSLOduration=2.349577008 podStartE2EDuration="5.728131288s" podCreationTimestamp="2026-02-27 19:46:31 +0000 UTC" firstStartedPulling="2026-02-27 19:46:32.584371703 +0000 UTC m=+754.229241438" lastFinishedPulling="2026-02-27 19:46:35.962925983 +0000 UTC m=+757.607795718" observedRunningTime="2026-02-27 19:46:36.72669946 +0000 UTC m=+758.371569235" watchObservedRunningTime="2026-02-27 19:46:36.728131288 +0000 UTC m=+758.373001023" Feb 27 19:46:36 crc kubenswrapper[4839]: I0227 19:46:36.728526 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" podStartSLOduration=2.224156908 podStartE2EDuration="5.728521748s" podCreationTimestamp="2026-02-27 19:46:31 +0000 UTC" firstStartedPulling="2026-02-27 19:46:32.405290856 +0000 UTC m=+754.050160591" lastFinishedPulling="2026-02-27 19:46:35.909655696 +0000 UTC m=+757.554525431" observedRunningTime="2026-02-27 19:46:36.712336375 +0000 UTC m=+758.357206130" watchObservedRunningTime="2026-02-27 19:46:36.728521748 +0000 UTC m=+758.373391483" Feb 27 19:46:38 crc kubenswrapper[4839]: E0227 19:46:38.970388 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:46:40 crc kubenswrapper[4839]: E0227 19:46:40.850260 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:46:40 crc kubenswrapper[4839]: E0227 19:46:40.850854 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:46:40 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:46:40 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m46pl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537026-67hjl_openshift-infra(9b5f3e68-002d-4f8f-a93a-b4854e58fc5b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:46:40 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:46:40 crc kubenswrapper[4839]: E0227 19:46:40.852058 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.101879 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t22j4"] Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102405 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-controller" containerID="cri-o://10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102473 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="nbdb" containerID="cri-o://ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102556 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="sbdb" containerID="cri-o://12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102579 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102557 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="northd" containerID="cri-o://616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102651 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-node" containerID="cri-o://31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.102660 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-acl-logging" containerID="cri-o://4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.193372 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" containerID="cri-o://58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" gracePeriod=30 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.194771 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-gxszm" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.439459 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/3.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.441615 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovn-acl-logging/0.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.442086 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovn-controller/0.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.442445 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487559 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hj2hr"] Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487779 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487793 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487802 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-acl-logging" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487808 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-acl-logging" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487817 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kubecfg-setup" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487824 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kubecfg-setup" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487834 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487841 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487851 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="northd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487857 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="northd" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487864 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="nbdb" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487869 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="nbdb" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487878 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487884 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487893 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487900 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487908 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487914 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487921 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487926 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487935 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-node" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487941 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-node" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.487948 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="sbdb" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.487954 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="sbdb" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488041 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488051 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="sbdb" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488059 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488066 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488075 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488082 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="kube-rbac-proxy-node" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488090 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="nbdb" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488097 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488103 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488111 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="northd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488119 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovn-acl-logging" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.488199 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488206 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.488301 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerName="ovnkube-controller" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.489893 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.580896 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-systemd-units\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581021 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-config\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581031 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581058 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68hlp\" (UniqueName: \"kubernetes.io/projected/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-kube-api-access-68hlp\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581147 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-node-log\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581210 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-systemd\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581208 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-node-log" (OuterVolumeSpecName: "node-log") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581275 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-script-lib\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581297 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-openvswitch\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581320 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-log-socket\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581382 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-kubelet\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581394 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581456 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-var-lib-openvswitch\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581481 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-log-socket" (OuterVolumeSpecName: "log-socket") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581485 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-ovn\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581510 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581509 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581505 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581516 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581553 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-bin\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581580 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-etc-openvswitch\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581620 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581631 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581634 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-slash\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581705 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-slash" (OuterVolumeSpecName: "host-slash") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581708 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-env-overrides\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581784 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-netns\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581812 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-ovn-kubernetes\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581853 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581855 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581867 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581884 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581901 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-netd\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581909 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581958 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.581961 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovn-node-metrics-cert\") pod \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\" (UID: \"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323\") " Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582036 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582369 4839 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582383 4839 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582393 4839 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-node-log\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582402 4839 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582410 4839 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582418 4839 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-log-socket\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582426 4839 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582435 4839 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582444 4839 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582452 4839 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582460 4839 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582468 4839 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-slash\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582478 4839 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582487 4839 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582497 4839 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582509 4839 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.582521 4839 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.587179 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-kube-api-access-68hlp" (OuterVolumeSpecName: "kube-api-access-68hlp") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "kube-api-access-68hlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.587202 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.593597 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" (UID: "1f65f341-ed5e-4c1c-8e7c-9c9e556c1323"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683678 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-node-log\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683727 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683750 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683766 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovnkube-script-lib\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683828 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-var-lib-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683890 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-cni-bin\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683928 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8f95\" (UniqueName: \"kubernetes.io/projected/d7bc8c88-560d-4ece-ba48-682e10a193bf-kube-api-access-c8f95\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683960 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-etc-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.683994 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-env-overrides\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684019 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-ovn\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684043 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-slash\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684056 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovn-node-metrics-cert\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684075 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-kubelet\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684093 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-run-netns\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684139 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-systemd-units\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684187 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovnkube-config\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684206 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-log-socket\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684221 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-cni-netd\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684236 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-run-ovn-kubernetes\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684256 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-systemd\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684317 4839 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684346 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68hlp\" (UniqueName: \"kubernetes.io/projected/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-kube-api-access-68hlp\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.684361 4839 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.730542 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/2.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.731014 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/1.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.731054 4839 generic.go:334] "Generic (PLEG): container finished" podID="050a1135-8d07-4d7c-89fe-f95d8b1e69f4" containerID="818928ceb7d6dc64a37536581b2018af0ff97985fed3a856d9292c377d5a5d0a" exitCode=2 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.731112 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerDied","Data":"818928ceb7d6dc64a37536581b2018af0ff97985fed3a856d9292c377d5a5d0a"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.731142 4839 scope.go:117] "RemoveContainer" containerID="b9b4e42bfe5dd2937a42fdf9a3338c1bc0c34c26d60c53c5441f69f309683250" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.731819 4839 scope.go:117] "RemoveContainer" containerID="818928ceb7d6dc64a37536581b2018af0ff97985fed3a856d9292c377d5a5d0a" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.732086 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2w9pp_openshift-multus(050a1135-8d07-4d7c-89fe-f95d8b1e69f4)\"" pod="openshift-multus/multus-2w9pp" podUID="050a1135-8d07-4d7c-89fe-f95d8b1e69f4" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.733784 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovnkube-controller/3.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.735537 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovn-acl-logging/0.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736023 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t22j4_1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/ovn-controller/0.log" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736472 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" exitCode=0 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736502 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" exitCode=0 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736512 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" exitCode=0 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736523 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" exitCode=0 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736532 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" exitCode=0 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736539 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" exitCode=0 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736545 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" exitCode=143 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736552 4839 generic.go:334] "Generic (PLEG): container finished" podID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" exitCode=143 Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736553 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736565 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736589 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736599 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736612 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736622 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736631 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736642 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736652 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736658 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736684 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736692 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736697 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736702 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736708 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736714 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736719 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736728 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736738 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736744 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736749 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736754 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736760 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736765 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736770 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736774 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736779 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736784 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736790 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736797 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736804 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736810 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736815 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736820 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736825 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736830 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736835 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736840 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736844 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736851 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t22j4" event={"ID":"1f65f341-ed5e-4c1c-8e7c-9c9e556c1323","Type":"ContainerDied","Data":"4bfa4dc5110c35782133a32333727c281789874a34ac1d94fc3fb67b3d182de1"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736860 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736866 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736871 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736877 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736881 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736886 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736892 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736897 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736901 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.736906 4839 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.756237 4839 scope.go:117] "RemoveContainer" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.782481 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t22j4"] Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.786050 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t22j4"] Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791177 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791406 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-kubelet\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791458 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-run-netns\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791516 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-kubelet\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791521 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-systemd-units\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791580 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovnkube-config\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791588 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-systemd-units\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791609 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-log-socket\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791611 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-run-netns\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791630 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-log-socket\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791653 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-cni-netd\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791690 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-cni-netd\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791716 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-run-ovn-kubernetes\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791768 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-systemd\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791788 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-run-ovn-kubernetes\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791807 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-node-log\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791819 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-systemd\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791886 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791935 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovnkube-script-lib\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791953 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.791972 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-var-lib-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792006 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792019 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-cni-bin\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792034 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-node-log\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792118 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792173 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-var-lib-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792209 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8f95\" (UniqueName: \"kubernetes.io/projected/d7bc8c88-560d-4ece-ba48-682e10a193bf-kube-api-access-c8f95\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792267 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-etc-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792335 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-env-overrides\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792363 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-ovn\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792382 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-slash\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792416 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovn-node-metrics-cert\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792340 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-cni-bin\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792597 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovnkube-script-lib\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792692 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-etc-openvswitch\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792794 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovnkube-config\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792886 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-run-ovn\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792941 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d7bc8c88-560d-4ece-ba48-682e10a193bf-host-slash\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.792962 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d7bc8c88-560d-4ece-ba48-682e10a193bf-env-overrides\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.796415 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d7bc8c88-560d-4ece-ba48-682e10a193bf-ovn-node-metrics-cert\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.809012 4839 scope.go:117] "RemoveContainer" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.811618 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8f95\" (UniqueName: \"kubernetes.io/projected/d7bc8c88-560d-4ece-ba48-682e10a193bf-kube-api-access-c8f95\") pod \"ovnkube-node-hj2hr\" (UID: \"d7bc8c88-560d-4ece-ba48-682e10a193bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.825112 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.826869 4839 scope.go:117] "RemoveContainer" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.847051 4839 scope.go:117] "RemoveContainer" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.863013 4839 scope.go:117] "RemoveContainer" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.884383 4839 scope.go:117] "RemoveContainer" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.900454 4839 scope.go:117] "RemoveContainer" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.926065 4839 scope.go:117] "RemoveContainer" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.948884 4839 scope.go:117] "RemoveContainer" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.977570 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f65f341-ed5e-4c1c-8e7c-9c9e556c1323" path="/var/lib/kubelet/pods/1f65f341-ed5e-4c1c-8e7c-9c9e556c1323/volumes" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.977650 4839 scope.go:117] "RemoveContainer" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.980402 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": container with ID starting with 58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d not found: ID does not exist" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.980453 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} err="failed to get container status \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": rpc error: code = NotFound desc = could not find container \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": container with ID starting with 58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.980481 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.981322 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": container with ID starting with 60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd not found: ID does not exist" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.981351 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} err="failed to get container status \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": rpc error: code = NotFound desc = could not find container \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": container with ID starting with 60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.981398 4839 scope.go:117] "RemoveContainer" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.981680 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": container with ID starting with 12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6 not found: ID does not exist" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.981716 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} err="failed to get container status \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": rpc error: code = NotFound desc = could not find container \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": container with ID starting with 12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.981741 4839 scope.go:117] "RemoveContainer" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.982033 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": container with ID starting with ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0 not found: ID does not exist" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.982057 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} err="failed to get container status \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": rpc error: code = NotFound desc = could not find container \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": container with ID starting with ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.982074 4839 scope.go:117] "RemoveContainer" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.982531 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": container with ID starting with 616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e not found: ID does not exist" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.982555 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} err="failed to get container status \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": rpc error: code = NotFound desc = could not find container \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": container with ID starting with 616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.982571 4839 scope.go:117] "RemoveContainer" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.982913 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": container with ID starting with bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880 not found: ID does not exist" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.982958 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} err="failed to get container status \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": rpc error: code = NotFound desc = could not find container \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": container with ID starting with bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.982993 4839 scope.go:117] "RemoveContainer" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.983287 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": container with ID starting with 31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701 not found: ID does not exist" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.983311 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} err="failed to get container status \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": rpc error: code = NotFound desc = could not find container \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": container with ID starting with 31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.983327 4839 scope.go:117] "RemoveContainer" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.983610 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": container with ID starting with 4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d not found: ID does not exist" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.983650 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} err="failed to get container status \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": rpc error: code = NotFound desc = could not find container \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": container with ID starting with 4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.983695 4839 scope.go:117] "RemoveContainer" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.983945 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": container with ID starting with 10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc not found: ID does not exist" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.983990 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} err="failed to get container status \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": rpc error: code = NotFound desc = could not find container \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": container with ID starting with 10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984013 4839 scope.go:117] "RemoveContainer" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" Feb 27 19:46:42 crc kubenswrapper[4839]: E0227 19:46:42.984290 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": container with ID starting with ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed not found: ID does not exist" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984315 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} err="failed to get container status \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": rpc error: code = NotFound desc = could not find container \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": container with ID starting with ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984332 4839 scope.go:117] "RemoveContainer" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984570 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} err="failed to get container status \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": rpc error: code = NotFound desc = could not find container \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": container with ID starting with 58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984591 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984846 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} err="failed to get container status \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": rpc error: code = NotFound desc = could not find container \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": container with ID starting with 60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.984874 4839 scope.go:117] "RemoveContainer" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.985230 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} err="failed to get container status \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": rpc error: code = NotFound desc = could not find container \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": container with ID starting with 12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.985254 4839 scope.go:117] "RemoveContainer" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.985502 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} err="failed to get container status \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": rpc error: code = NotFound desc = could not find container \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": container with ID starting with ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.985525 4839 scope.go:117] "RemoveContainer" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.985842 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} err="failed to get container status \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": rpc error: code = NotFound desc = could not find container \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": container with ID starting with 616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.985868 4839 scope.go:117] "RemoveContainer" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.986129 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} err="failed to get container status \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": rpc error: code = NotFound desc = could not find container \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": container with ID starting with bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.986155 4839 scope.go:117] "RemoveContainer" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.986861 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} err="failed to get container status \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": rpc error: code = NotFound desc = could not find container \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": container with ID starting with 31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.986891 4839 scope.go:117] "RemoveContainer" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.987136 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} err="failed to get container status \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": rpc error: code = NotFound desc = could not find container \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": container with ID starting with 4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.987165 4839 scope.go:117] "RemoveContainer" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.987367 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} err="failed to get container status \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": rpc error: code = NotFound desc = could not find container \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": container with ID starting with 10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.987388 4839 scope.go:117] "RemoveContainer" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.988106 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} err="failed to get container status \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": rpc error: code = NotFound desc = could not find container \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": container with ID starting with ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.988136 4839 scope.go:117] "RemoveContainer" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.988461 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} err="failed to get container status \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": rpc error: code = NotFound desc = could not find container \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": container with ID starting with 58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.988481 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.988823 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} err="failed to get container status \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": rpc error: code = NotFound desc = could not find container \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": container with ID starting with 60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.988842 4839 scope.go:117] "RemoveContainer" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.989578 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} err="failed to get container status \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": rpc error: code = NotFound desc = could not find container \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": container with ID starting with 12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.989612 4839 scope.go:117] "RemoveContainer" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.989905 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} err="failed to get container status \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": rpc error: code = NotFound desc = could not find container \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": container with ID starting with ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.989932 4839 scope.go:117] "RemoveContainer" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.990269 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} err="failed to get container status \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": rpc error: code = NotFound desc = could not find container \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": container with ID starting with 616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.990296 4839 scope.go:117] "RemoveContainer" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.990613 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} err="failed to get container status \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": rpc error: code = NotFound desc = could not find container \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": container with ID starting with bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.990636 4839 scope.go:117] "RemoveContainer" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.990924 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} err="failed to get container status \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": rpc error: code = NotFound desc = could not find container \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": container with ID starting with 31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.990954 4839 scope.go:117] "RemoveContainer" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.991262 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} err="failed to get container status \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": rpc error: code = NotFound desc = could not find container \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": container with ID starting with 4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.991295 4839 scope.go:117] "RemoveContainer" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.991532 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} err="failed to get container status \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": rpc error: code = NotFound desc = could not find container \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": container with ID starting with 10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.991577 4839 scope.go:117] "RemoveContainer" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.991869 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} err="failed to get container status \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": rpc error: code = NotFound desc = could not find container \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": container with ID starting with ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.991890 4839 scope.go:117] "RemoveContainer" containerID="58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.992163 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d"} err="failed to get container status \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": rpc error: code = NotFound desc = could not find container \"58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d\": container with ID starting with 58d4eff2731cd86094f800946643517cd65c7dfcc7e3cf06f96cadbbcf428c3d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.992195 4839 scope.go:117] "RemoveContainer" containerID="60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.992492 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd"} err="failed to get container status \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": rpc error: code = NotFound desc = could not find container \"60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd\": container with ID starting with 60116770421b0d3a86743bd1435ea1808b8e38fbdb30f694f43e8acadf532bfd not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.992520 4839 scope.go:117] "RemoveContainer" containerID="12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.992842 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6"} err="failed to get container status \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": rpc error: code = NotFound desc = could not find container \"12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6\": container with ID starting with 12549656fa2e9c7ef62876ebea9df13906b88ec004bfe617d2ae6343570c1da6 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.992873 4839 scope.go:117] "RemoveContainer" containerID="ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.993130 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0"} err="failed to get container status \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": rpc error: code = NotFound desc = could not find container \"ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0\": container with ID starting with ff30d66fa2f191e2f61d197455154dc24b1b8bb2c3f6eb7c3fd0fd813f6838c0 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.993147 4839 scope.go:117] "RemoveContainer" containerID="616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.993407 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e"} err="failed to get container status \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": rpc error: code = NotFound desc = could not find container \"616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e\": container with ID starting with 616dcfbecabaef1cd37c0e6f6714773e9ec9f4ad725e3e5995b193e5b81a562e not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.993428 4839 scope.go:117] "RemoveContainer" containerID="bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.994928 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880"} err="failed to get container status \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": rpc error: code = NotFound desc = could not find container \"bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880\": container with ID starting with bda8345e2e5fd5f3f983d883475eb4996ec88955279d08744ce6411e25475880 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.995230 4839 scope.go:117] "RemoveContainer" containerID="31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.995531 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701"} err="failed to get container status \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": rpc error: code = NotFound desc = could not find container \"31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701\": container with ID starting with 31a05dc1953c281825c4ce2668c2d0b44e5aa35f4178625af2c2be2387db4701 not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.995558 4839 scope.go:117] "RemoveContainer" containerID="4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.995915 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d"} err="failed to get container status \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": rpc error: code = NotFound desc = could not find container \"4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d\": container with ID starting with 4f267965f86b9597186355813ab7862e082dbcecec6326864823b2ddf5b87d0d not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.995945 4839 scope.go:117] "RemoveContainer" containerID="10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.996218 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc"} err="failed to get container status \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": rpc error: code = NotFound desc = could not find container \"10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc\": container with ID starting with 10f96a43974353c592626d9f370aebdeba5ca386a58b3aab39534027799c10dc not found: ID does not exist" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.996264 4839 scope.go:117] "RemoveContainer" containerID="ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed" Feb 27 19:46:42 crc kubenswrapper[4839]: I0227 19:46:42.996536 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed"} err="failed to get container status \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": rpc error: code = NotFound desc = could not find container \"ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed\": container with ID starting with ea4066670e4f5a4d90fd26c638bf71a6cedfceab979726e12c088c995347d9ed not found: ID does not exist" Feb 27 19:46:43 crc kubenswrapper[4839]: I0227 19:46:43.746433 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/2.log" Feb 27 19:46:43 crc kubenswrapper[4839]: I0227 19:46:43.749611 4839 generic.go:334] "Generic (PLEG): container finished" podID="d7bc8c88-560d-4ece-ba48-682e10a193bf" containerID="4aa69ecd732b194b3640d30eefac6d68060bf5e3ed2ed33953b02370291de045" exitCode=0 Feb 27 19:46:43 crc kubenswrapper[4839]: I0227 19:46:43.749682 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerDied","Data":"4aa69ecd732b194b3640d30eefac6d68060bf5e3ed2ed33953b02370291de045"} Feb 27 19:46:43 crc kubenswrapper[4839]: I0227 19:46:43.749707 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"46bdf9ab21c149365875470ccb7e7582ee6ddf1b33b33e58290699a2c502fca0"} Feb 27 19:46:44 crc kubenswrapper[4839]: I0227 19:46:44.775420 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"0b310ab72caa4a481b6bda76663b116d45ac3e76c94db2f43f6e2eb526b9e077"} Feb 27 19:46:44 crc kubenswrapper[4839]: I0227 19:46:44.775933 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"da943a2fbe1e8940021032154706eea5eaff036a2075654a913f81accab6c5e9"} Feb 27 19:46:44 crc kubenswrapper[4839]: I0227 19:46:44.775960 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"a806133ee19da1980ad0020cdaab4f800f49f19473f23501217dafd4a2e63373"} Feb 27 19:46:44 crc kubenswrapper[4839]: I0227 19:46:44.775979 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"a7e1ea1150b0bf022ddc23eccab8e3336d900947f43f106d13d630b3c9207efd"} Feb 27 19:46:44 crc kubenswrapper[4839]: I0227 19:46:44.775996 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"b2e98b47948ddd6b289c610849d6129f4c88a62bc697b5752b8ab07eab8834c0"} Feb 27 19:46:44 crc kubenswrapper[4839]: I0227 19:46:44.776013 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"41752caf4836d496e5afb1e274446f9a30bfda3ca10789984a4c436f70f4f70a"} Feb 27 19:46:45 crc kubenswrapper[4839]: E0227 19:46:45.967529 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:46:47 crc kubenswrapper[4839]: I0227 19:46:47.805770 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"4458ac46429ea42a89e1a5c33f819ddc66808bf04987c7c7ec7320fc40629f6a"} Feb 27 19:46:49 crc kubenswrapper[4839]: I0227 19:46:49.823714 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" event={"ID":"d7bc8c88-560d-4ece-ba48-682e10a193bf","Type":"ContainerStarted","Data":"d749c7e691bb38ea63a2e8ddbbe7dc97c02cb9902db12e11df4b67a623db04b4"} Feb 27 19:46:49 crc kubenswrapper[4839]: I0227 19:46:49.824113 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:49 crc kubenswrapper[4839]: I0227 19:46:49.824124 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:49 crc kubenswrapper[4839]: I0227 19:46:49.861144 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:49 crc kubenswrapper[4839]: I0227 19:46:49.861836 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" podStartSLOduration=7.861818525 podStartE2EDuration="7.861818525s" podCreationTimestamp="2026-02-27 19:46:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 19:46:49.859193714 +0000 UTC m=+771.504063479" watchObservedRunningTime="2026-02-27 19:46:49.861818525 +0000 UTC m=+771.506688260" Feb 27 19:46:50 crc kubenswrapper[4839]: I0227 19:46:50.831104 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:50 crc kubenswrapper[4839]: I0227 19:46:50.891819 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:46:50 crc kubenswrapper[4839]: E0227 19:46:50.967864 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:46:51 crc kubenswrapper[4839]: E0227 19:46:51.966939 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:46:52 crc kubenswrapper[4839]: I0227 19:46:52.682144 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:46:52 crc kubenswrapper[4839]: I0227 19:46:52.682241 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:46:54 crc kubenswrapper[4839]: I0227 19:46:54.966421 4839 scope.go:117] "RemoveContainer" containerID="818928ceb7d6dc64a37536581b2018af0ff97985fed3a856d9292c377d5a5d0a" Feb 27 19:46:54 crc kubenswrapper[4839]: E0227 19:46:54.967090 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2w9pp_openshift-multus(050a1135-8d07-4d7c-89fe-f95d8b1e69f4)\"" pod="openshift-multus/multus-2w9pp" podUID="050a1135-8d07-4d7c-89fe-f95d8b1e69f4" Feb 27 19:47:00 crc kubenswrapper[4839]: E0227 19:47:00.855835 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:47:00 crc kubenswrapper[4839]: E0227 19:47:00.856578 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:47:00 crc kubenswrapper[4839]: E0227 19:47:00.857881 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:47:05 crc kubenswrapper[4839]: I0227 19:47:05.965416 4839 scope.go:117] "RemoveContainer" containerID="818928ceb7d6dc64a37536581b2018af0ff97985fed3a856d9292c377d5a5d0a" Feb 27 19:47:05 crc kubenswrapper[4839]: E0227 19:47:05.968758 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:47:06 crc kubenswrapper[4839]: E0227 19:47:06.333329 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:47:06 crc kubenswrapper[4839]: E0227 19:47:06.333836 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:47:06 crc kubenswrapper[4839]: E0227 19:47:06.335046 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:47:06 crc kubenswrapper[4839]: I0227 19:47:06.943760 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2w9pp_050a1135-8d07-4d7c-89fe-f95d8b1e69f4/kube-multus/2.log" Feb 27 19:47:06 crc kubenswrapper[4839]: I0227 19:47:06.944101 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2w9pp" event={"ID":"050a1135-8d07-4d7c-89fe-f95d8b1e69f4","Type":"ContainerStarted","Data":"ffcbd0312d6b52d6b2256b63818e142feaa888212dd0b87d92d3586a6b1b81e8"} Feb 27 19:47:12 crc kubenswrapper[4839]: I0227 19:47:12.852038 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hj2hr" Feb 27 19:47:13 crc kubenswrapper[4839]: E0227 19:47:13.969880 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:47:17 crc kubenswrapper[4839]: E0227 19:47:17.969747 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:47:19 crc kubenswrapper[4839]: E0227 19:47:19.967111 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:47:22 crc kubenswrapper[4839]: I0227 19:47:22.682793 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:47:22 crc kubenswrapper[4839]: I0227 19:47:22.683358 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:47:24 crc kubenswrapper[4839]: E0227 19:47:24.970065 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:47:30 crc kubenswrapper[4839]: E0227 19:47:30.146283 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:47:30 crc kubenswrapper[4839]: E0227 19:47:30.147025 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:47:30 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:47:30 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m46pl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537026-67hjl_openshift-infra(9b5f3e68-002d-4f8f-a93a-b4854e58fc5b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:47:30 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:47:30 crc kubenswrapper[4839]: E0227 19:47:30.148370 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:47:34 crc kubenswrapper[4839]: E0227 19:47:34.969713 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.038825 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7hm7c"] Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.040387 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.062149 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hm7c"] Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.226359 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-catalog-content\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.226435 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzh9q\" (UniqueName: \"kubernetes.io/projected/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-kube-api-access-zzh9q\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.226485 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-utilities\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.327576 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzh9q\" (UniqueName: \"kubernetes.io/projected/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-kube-api-access-zzh9q\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.327736 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-utilities\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.327822 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-catalog-content\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.328266 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-utilities\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.328507 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-catalog-content\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.367183 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzh9q\" (UniqueName: \"kubernetes.io/projected/02ff6284-1eb0-403a-8bfd-67184d2b3f5e-kube-api-access-zzh9q\") pod \"redhat-marketplace-7hm7c\" (UID: \"02ff6284-1eb0-403a-8bfd-67184d2b3f5e\") " pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.662508 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hm7c" Feb 27 19:47:35 crc kubenswrapper[4839]: I0227 19:47:35.880243 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hm7c"] Feb 27 19:47:35 crc kubenswrapper[4839]: E0227 19:47:35.967749 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:47:36 crc kubenswrapper[4839]: I0227 19:47:36.152529 4839 generic.go:334] "Generic (PLEG): container finished" podID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" containerID="465bdc68a379e902aa3950285269b9ce30a039443b060c3a54a8961de5caaf66" exitCode=0 Feb 27 19:47:36 crc kubenswrapper[4839]: I0227 19:47:36.152570 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hm7c" event={"ID":"02ff6284-1eb0-403a-8bfd-67184d2b3f5e","Type":"ContainerDied","Data":"465bdc68a379e902aa3950285269b9ce30a039443b060c3a54a8961de5caaf66"} Feb 27 19:47:36 crc kubenswrapper[4839]: I0227 19:47:36.152594 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hm7c" event={"ID":"02ff6284-1eb0-403a-8bfd-67184d2b3f5e","Type":"ContainerStarted","Data":"33c881abe1d1d90458495c44e9457b7bd2d7f9d2669c492a756eb65925c925a9"} Feb 27 19:47:36 crc kubenswrapper[4839]: E0227 19:47:36.680543 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:47:36 crc kubenswrapper[4839]: E0227 19:47:36.681171 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzh9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7hm7c_openshift-marketplace(02ff6284-1eb0-403a-8bfd-67184d2b3f5e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:47:36 crc kubenswrapper[4839]: E0227 19:47:36.682430 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:47:37 crc kubenswrapper[4839]: E0227 19:47:37.159777 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:47:38 crc kubenswrapper[4839]: I0227 19:47:38.479529 4839 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 19:47:40 crc kubenswrapper[4839]: E0227 19:47:40.969332 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:47:47 crc kubenswrapper[4839]: E0227 19:47:47.969171 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:47:47 crc kubenswrapper[4839]: I0227 19:47:47.969254 4839 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 19:47:48 crc kubenswrapper[4839]: E0227 19:47:48.625945 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:47:48 crc kubenswrapper[4839]: E0227 19:47:48.626506 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzh9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7hm7c_openshift-marketplace(02ff6284-1eb0-403a-8bfd-67184d2b3f5e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:47:48 crc kubenswrapper[4839]: E0227 19:47:48.628369 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:47:48 crc kubenswrapper[4839]: E0227 19:47:48.970481 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:47:52 crc kubenswrapper[4839]: I0227 19:47:52.681975 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:47:52 crc kubenswrapper[4839]: I0227 19:47:52.682283 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:47:52 crc kubenswrapper[4839]: I0227 19:47:52.682327 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:47:52 crc kubenswrapper[4839]: I0227 19:47:52.682841 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fc60b748d670020094881ec089d53e0f67db6413e9b475783fdc7432a738c18"} pod="openshift-machine-config-operator/machine-config-daemon-thb8n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 19:47:52 crc kubenswrapper[4839]: I0227 19:47:52.682894 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" containerID="cri-o://5fc60b748d670020094881ec089d53e0f67db6413e9b475783fdc7432a738c18" gracePeriod=600 Feb 27 19:47:53 crc kubenswrapper[4839]: I0227 19:47:53.259402 4839 generic.go:334] "Generic (PLEG): container finished" podID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerID="5fc60b748d670020094881ec089d53e0f67db6413e9b475783fdc7432a738c18" exitCode=0 Feb 27 19:47:53 crc kubenswrapper[4839]: I0227 19:47:53.259491 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerDied","Data":"5fc60b748d670020094881ec089d53e0f67db6413e9b475783fdc7432a738c18"} Feb 27 19:47:53 crc kubenswrapper[4839]: I0227 19:47:53.259758 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"cebc994c5b008539c690e93a6330de8ce7ecd9ea0602f50a6d17fdfe769f11e8"} Feb 27 19:47:53 crc kubenswrapper[4839]: I0227 19:47:53.259786 4839 scope.go:117] "RemoveContainer" containerID="1b68014096a9cd142c188dd5e27a95973d3809ea3da725eb9ba31acfc05d2d48" Feb 27 19:47:54 crc kubenswrapper[4839]: E0227 19:47:54.967508 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.156620 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537028-zkw8j"] Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.158473 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.165963 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537028-zkw8j"] Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.166982 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrzf\" (UniqueName: \"kubernetes.io/projected/5f8c00af-e3ff-439e-b97f-83b225d5a234-kube-api-access-hgrzf\") pod \"auto-csr-approver-29537028-zkw8j\" (UID: \"5f8c00af-e3ff-439e-b97f-83b225d5a234\") " pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.268139 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrzf\" (UniqueName: \"kubernetes.io/projected/5f8c00af-e3ff-439e-b97f-83b225d5a234-kube-api-access-hgrzf\") pod \"auto-csr-approver-29537028-zkw8j\" (UID: \"5f8c00af-e3ff-439e-b97f-83b225d5a234\") " pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.292846 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrzf\" (UniqueName: \"kubernetes.io/projected/5f8c00af-e3ff-439e-b97f-83b225d5a234-kube-api-access-hgrzf\") pod \"auto-csr-approver-29537028-zkw8j\" (UID: \"5f8c00af-e3ff-439e-b97f-83b225d5a234\") " pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.495832 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:48:00 crc kubenswrapper[4839]: I0227 19:48:00.752482 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537028-zkw8j"] Feb 27 19:48:00 crc kubenswrapper[4839]: W0227 19:48:00.760909 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f8c00af_e3ff_439e_b97f_83b225d5a234.slice/crio-644bd59fd64472453806caa2b06788a97695a17b7936cfa3a2af62c56bdbefe9 WatchSource:0}: Error finding container 644bd59fd64472453806caa2b06788a97695a17b7936cfa3a2af62c56bdbefe9: Status 404 returned error can't find the container with id 644bd59fd64472453806caa2b06788a97695a17b7936cfa3a2af62c56bdbefe9 Feb 27 19:48:00 crc kubenswrapper[4839]: E0227 19:48:00.968348 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:48:01 crc kubenswrapper[4839]: I0227 19:48:01.320454 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" event={"ID":"5f8c00af-e3ff-439e-b97f-83b225d5a234","Type":"ContainerStarted","Data":"644bd59fd64472453806caa2b06788a97695a17b7936cfa3a2af62c56bdbefe9"} Feb 27 19:48:01 crc kubenswrapper[4839]: E0227 19:48:01.970527 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:48:02 crc kubenswrapper[4839]: E0227 19:48:02.020998 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:48:02 crc kubenswrapper[4839]: E0227 19:48:02.021311 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:48:02 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:48:02 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hgrzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537028-zkw8j_openshift-infra(5f8c00af-e3ff-439e-b97f-83b225d5a234): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:48:02 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:48:02 crc kubenswrapper[4839]: E0227 19:48:02.022739 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:48:02 crc kubenswrapper[4839]: E0227 19:48:02.328309 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:48:02 crc kubenswrapper[4839]: E0227 19:48:02.967637 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:48:08 crc kubenswrapper[4839]: E0227 19:48:08.971710 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:48:12 crc kubenswrapper[4839]: E0227 19:48:12.969348 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:48:14 crc kubenswrapper[4839]: E0227 19:48:14.066892 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:48:14 crc kubenswrapper[4839]: E0227 19:48:14.067072 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzh9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7hm7c_openshift-marketplace(02ff6284-1eb0-403a-8bfd-67184d2b3f5e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:48:14 crc kubenswrapper[4839]: E0227 19:48:14.068314 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:48:15 crc kubenswrapper[4839]: E0227 19:48:15.883572 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:48:15 crc kubenswrapper[4839]: E0227 19:48:15.884910 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:48:15 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:48:15 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hgrzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537028-zkw8j_openshift-infra(5f8c00af-e3ff-439e-b97f-83b225d5a234): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:48:15 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:48:15 crc kubenswrapper[4839]: E0227 19:48:15.886493 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:48:15 crc kubenswrapper[4839]: E0227 19:48:15.968023 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:48:21 crc kubenswrapper[4839]: E0227 19:48:21.967939 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:48:24 crc kubenswrapper[4839]: E0227 19:48:24.969040 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:48:26 crc kubenswrapper[4839]: E0227 19:48:26.968227 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:48:28 crc kubenswrapper[4839]: E0227 19:48:28.969911 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:48:28 crc kubenswrapper[4839]: E0227 19:48:28.970315 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:48:35 crc kubenswrapper[4839]: E0227 19:48:35.968932 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:48:37 crc kubenswrapper[4839]: E0227 19:48:37.967113 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:48:39 crc kubenswrapper[4839]: E0227 19:48:39.967933 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:48:39 crc kubenswrapper[4839]: E0227 19:48:39.968312 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:48:42 crc kubenswrapper[4839]: E0227 19:48:42.801517 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:48:42 crc kubenswrapper[4839]: E0227 19:48:42.801799 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:48:42 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:48:42 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hgrzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537028-zkw8j_openshift-infra(5f8c00af-e3ff-439e-b97f-83b225d5a234): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:48:42 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:48:42 crc kubenswrapper[4839]: E0227 19:48:42.803082 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:48:47 crc kubenswrapper[4839]: E0227 19:48:47.967470 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537026-67hjl" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" Feb 27 19:48:50 crc kubenswrapper[4839]: E0227 19:48:50.969451 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:48:51 crc kubenswrapper[4839]: E0227 19:48:51.967169 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:48:52 crc kubenswrapper[4839]: E0227 19:48:52.968108 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:48:54 crc kubenswrapper[4839]: E0227 19:48:54.966785 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:49:02 crc kubenswrapper[4839]: E0227 19:49:02.966522 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:49:03 crc kubenswrapper[4839]: I0227 19:49:03.735331 4839 generic.go:334] "Generic (PLEG): container finished" podID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" containerID="d6c064dd8724db26eb48b2a63397b4770a988c3ff9f288dad5702756f9db3b1a" exitCode=0 Feb 27 19:49:03 crc kubenswrapper[4839]: I0227 19:49:03.735559 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537026-67hjl" event={"ID":"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b","Type":"ContainerDied","Data":"d6c064dd8724db26eb48b2a63397b4770a988c3ff9f288dad5702756f9db3b1a"} Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.045094 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.244485 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m46pl\" (UniqueName: \"kubernetes.io/projected/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b-kube-api-access-m46pl\") pod \"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b\" (UID: \"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b\") " Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.252619 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b-kube-api-access-m46pl" (OuterVolumeSpecName: "kube-api-access-m46pl") pod "9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" (UID: "9b5f3e68-002d-4f8f-a93a-b4854e58fc5b"). InnerVolumeSpecName "kube-api-access-m46pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.346361 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m46pl\" (UniqueName: \"kubernetes.io/projected/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b-kube-api-access-m46pl\") on node \"crc\" DevicePath \"\"" Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.753193 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537026-67hjl" event={"ID":"9b5f3e68-002d-4f8f-a93a-b4854e58fc5b","Type":"ContainerDied","Data":"f9c3d38bdfbb7425d422f105b01a646cd638d71310d6bc0d9662782093c13d58"} Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.753241 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9c3d38bdfbb7425d422f105b01a646cd638d71310d6bc0d9662782093c13d58" Feb 27 19:49:05 crc kubenswrapper[4839]: I0227 19:49:05.753300 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537026-67hjl" Feb 27 19:49:05 crc kubenswrapper[4839]: E0227 19:49:05.966823 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:49:06 crc kubenswrapper[4839]: I0227 19:49:06.117855 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537020-pq8zb"] Feb 27 19:49:06 crc kubenswrapper[4839]: I0227 19:49:06.125869 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537020-pq8zb"] Feb 27 19:49:06 crc kubenswrapper[4839]: I0227 19:49:06.972750 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fa05b9a-2335-4621-9c59-147a735f526f" path="/var/lib/kubelet/pods/4fa05b9a-2335-4621-9c59-147a735f526f/volumes" Feb 27 19:49:07 crc kubenswrapper[4839]: E0227 19:49:07.967833 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:49:08 crc kubenswrapper[4839]: E0227 19:49:08.708606 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:49:08 crc kubenswrapper[4839]: E0227 19:49:08.708807 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzh9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7hm7c_openshift-marketplace(02ff6284-1eb0-403a-8bfd-67184d2b3f5e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:49:08 crc kubenswrapper[4839]: E0227 19:49:08.710092 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:49:15 crc kubenswrapper[4839]: E0227 19:49:15.968740 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:49:19 crc kubenswrapper[4839]: E0227 19:49:19.968849 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:49:21 crc kubenswrapper[4839]: E0227 19:49:21.967994 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" Feb 27 19:49:21 crc kubenswrapper[4839]: E0227 19:49:21.968287 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:49:26 crc kubenswrapper[4839]: E0227 19:49:26.968212 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.058102 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9j7z5"] Feb 27 19:49:27 crc kubenswrapper[4839]: E0227 19:49:27.059544 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" containerName="oc" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.059774 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" containerName="oc" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.060248 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" containerName="oc" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.063796 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.067424 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9j7z5"] Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.169621 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-utilities\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.169774 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-catalog-content\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.169801 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s7kt\" (UniqueName: \"kubernetes.io/projected/b64b6367-e435-4dac-aa22-ac7efafdcd21-kube-api-access-5s7kt\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.270578 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-utilities\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.270644 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-catalog-content\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.270687 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s7kt\" (UniqueName: \"kubernetes.io/projected/b64b6367-e435-4dac-aa22-ac7efafdcd21-kube-api-access-5s7kt\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.271559 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-utilities\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.271914 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-catalog-content\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.293186 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s7kt\" (UniqueName: \"kubernetes.io/projected/b64b6367-e435-4dac-aa22-ac7efafdcd21-kube-api-access-5s7kt\") pod \"redhat-operators-9j7z5\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.396573 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.832587 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9j7z5"] Feb 27 19:49:27 crc kubenswrapper[4839]: I0227 19:49:27.908197 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerStarted","Data":"83e3b6ceb67a93c006aa863a34483d4e75d5de16fc01d14d24d1cf2d8b00216b"} Feb 27 19:49:28 crc kubenswrapper[4839]: I0227 19:49:28.917136 4839 generic.go:334] "Generic (PLEG): container finished" podID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerID="86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127" exitCode=0 Feb 27 19:49:28 crc kubenswrapper[4839]: I0227 19:49:28.917179 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerDied","Data":"86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127"} Feb 27 19:49:29 crc kubenswrapper[4839]: E0227 19:49:29.740477 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:49:29 crc kubenswrapper[4839]: E0227 19:49:29.740930 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5s7kt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9j7z5_openshift-marketplace(b64b6367-e435-4dac-aa22-ac7efafdcd21): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:49:29 crc kubenswrapper[4839]: E0227 19:49:29.742115 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-9j7z5" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" Feb 27 19:49:29 crc kubenswrapper[4839]: E0227 19:49:29.923148 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9j7z5" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" Feb 27 19:49:30 crc kubenswrapper[4839]: E0227 19:49:30.967646 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:49:32 crc kubenswrapper[4839]: E0227 19:49:32.968372 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:49:37 crc kubenswrapper[4839]: E0227 19:49:37.967014 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:49:38 crc kubenswrapper[4839]: I0227 19:49:38.988961 4839 generic.go:334] "Generic (PLEG): container finished" podID="5f8c00af-e3ff-439e-b97f-83b225d5a234" containerID="c923cf593fa05b58168fe6145e61e21c647dbb090e03dcb78a49a6cf594d108e" exitCode=0 Feb 27 19:49:38 crc kubenswrapper[4839]: I0227 19:49:38.989042 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" event={"ID":"5f8c00af-e3ff-439e-b97f-83b225d5a234","Type":"ContainerDied","Data":"c923cf593fa05b58168fe6145e61e21c647dbb090e03dcb78a49a6cf594d108e"} Feb 27 19:49:40 crc kubenswrapper[4839]: I0227 19:49:40.223830 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:49:40 crc kubenswrapper[4839]: I0227 19:49:40.343076 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgrzf\" (UniqueName: \"kubernetes.io/projected/5f8c00af-e3ff-439e-b97f-83b225d5a234-kube-api-access-hgrzf\") pod \"5f8c00af-e3ff-439e-b97f-83b225d5a234\" (UID: \"5f8c00af-e3ff-439e-b97f-83b225d5a234\") " Feb 27 19:49:40 crc kubenswrapper[4839]: I0227 19:49:40.348638 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f8c00af-e3ff-439e-b97f-83b225d5a234-kube-api-access-hgrzf" (OuterVolumeSpecName: "kube-api-access-hgrzf") pod "5f8c00af-e3ff-439e-b97f-83b225d5a234" (UID: "5f8c00af-e3ff-439e-b97f-83b225d5a234"). InnerVolumeSpecName "kube-api-access-hgrzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:49:40 crc kubenswrapper[4839]: I0227 19:49:40.444579 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgrzf\" (UniqueName: \"kubernetes.io/projected/5f8c00af-e3ff-439e-b97f-83b225d5a234-kube-api-access-hgrzf\") on node \"crc\" DevicePath \"\"" Feb 27 19:49:41 crc kubenswrapper[4839]: I0227 19:49:41.002453 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" event={"ID":"5f8c00af-e3ff-439e-b97f-83b225d5a234","Type":"ContainerDied","Data":"644bd59fd64472453806caa2b06788a97695a17b7936cfa3a2af62c56bdbefe9"} Feb 27 19:49:41 crc kubenswrapper[4839]: I0227 19:49:41.002490 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="644bd59fd64472453806caa2b06788a97695a17b7936cfa3a2af62c56bdbefe9" Feb 27 19:49:41 crc kubenswrapper[4839]: I0227 19:49:41.002535 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537028-zkw8j" Feb 27 19:49:41 crc kubenswrapper[4839]: I0227 19:49:41.292837 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537022-qgfbs"] Feb 27 19:49:41 crc kubenswrapper[4839]: I0227 19:49:41.297548 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537022-qgfbs"] Feb 27 19:49:41 crc kubenswrapper[4839]: E0227 19:49:41.967399 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:49:42 crc kubenswrapper[4839]: E0227 19:49:42.589412 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:49:42 crc kubenswrapper[4839]: E0227 19:49:42.589896 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5s7kt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9j7z5_openshift-marketplace(b64b6367-e435-4dac-aa22-ac7efafdcd21): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:49:42 crc kubenswrapper[4839]: E0227 19:49:42.591251 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-9j7z5" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" Feb 27 19:49:42 crc kubenswrapper[4839]: I0227 19:49:42.977482 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d9c59c-867a-41d5-b435-7c0693c857cb" path="/var/lib/kubelet/pods/36d9c59c-867a-41d5-b435-7c0693c857cb/volumes" Feb 27 19:49:46 crc kubenswrapper[4839]: E0227 19:49:46.967591 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.149792 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-68nwh"] Feb 27 19:49:48 crc kubenswrapper[4839]: E0227 19:49:48.150506 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" containerName="oc" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.150533 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" containerName="oc" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.150740 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f8c00af-e3ff-439e-b97f-83b225d5a234" containerName="oc" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.153014 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.161106 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68nwh"] Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.348827 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-utilities\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.348865 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsb8d\" (UniqueName: \"kubernetes.io/projected/88a2d66b-54c2-401e-a822-1a166817b82e-kube-api-access-dsb8d\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.348924 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-catalog-content\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.450188 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-catalog-content\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.450344 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-utilities\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.450365 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsb8d\" (UniqueName: \"kubernetes.io/projected/88a2d66b-54c2-401e-a822-1a166817b82e-kube-api-access-dsb8d\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.450866 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-catalog-content\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.450879 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-utilities\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.471549 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsb8d\" (UniqueName: \"kubernetes.io/projected/88a2d66b-54c2-401e-a822-1a166817b82e-kube-api-access-dsb8d\") pod \"certified-operators-68nwh\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.507925 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:48 crc kubenswrapper[4839]: I0227 19:49:48.913994 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68nwh"] Feb 27 19:49:48 crc kubenswrapper[4839]: W0227 19:49:48.937824 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88a2d66b_54c2_401e_a822_1a166817b82e.slice/crio-c758200e6946203934c30750aa68c29134640f0593a4d03e31b9d43b1073826b WatchSource:0}: Error finding container c758200e6946203934c30750aa68c29134640f0593a4d03e31b9d43b1073826b: Status 404 returned error can't find the container with id c758200e6946203934c30750aa68c29134640f0593a4d03e31b9d43b1073826b Feb 27 19:49:48 crc kubenswrapper[4839]: E0227 19:49:48.979564 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:49:49 crc kubenswrapper[4839]: I0227 19:49:49.053940 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerStarted","Data":"c758200e6946203934c30750aa68c29134640f0593a4d03e31b9d43b1073826b"} Feb 27 19:49:50 crc kubenswrapper[4839]: I0227 19:49:50.060721 4839 generic.go:334] "Generic (PLEG): container finished" podID="88a2d66b-54c2-401e-a822-1a166817b82e" containerID="ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65" exitCode=0 Feb 27 19:49:50 crc kubenswrapper[4839]: I0227 19:49:50.060882 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerDied","Data":"ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65"} Feb 27 19:49:51 crc kubenswrapper[4839]: I0227 19:49:51.068041 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerStarted","Data":"5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203"} Feb 27 19:49:52 crc kubenswrapper[4839]: I0227 19:49:52.074645 4839 generic.go:334] "Generic (PLEG): container finished" podID="88a2d66b-54c2-401e-a822-1a166817b82e" containerID="5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203" exitCode=0 Feb 27 19:49:52 crc kubenswrapper[4839]: I0227 19:49:52.074766 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerDied","Data":"5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203"} Feb 27 19:49:52 crc kubenswrapper[4839]: I0227 19:49:52.682728 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:49:52 crc kubenswrapper[4839]: I0227 19:49:52.682814 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:49:53 crc kubenswrapper[4839]: I0227 19:49:53.085175 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerStarted","Data":"84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03"} Feb 27 19:49:53 crc kubenswrapper[4839]: I0227 19:49:53.118737 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-68nwh" podStartSLOduration=2.669116195 podStartE2EDuration="5.118704207s" podCreationTimestamp="2026-02-27 19:49:48 +0000 UTC" firstStartedPulling="2026-02-27 19:49:50.062533116 +0000 UTC m=+951.707402851" lastFinishedPulling="2026-02-27 19:49:52.512121088 +0000 UTC m=+954.156990863" observedRunningTime="2026-02-27 19:49:53.111822474 +0000 UTC m=+954.756692289" watchObservedRunningTime="2026-02-27 19:49:53.118704207 +0000 UTC m=+954.763573992" Feb 27 19:49:53 crc kubenswrapper[4839]: E0227 19:49:53.969747 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9j7z5" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" Feb 27 19:49:56 crc kubenswrapper[4839]: E0227 19:49:56.970147 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:49:58 crc kubenswrapper[4839]: I0227 19:49:58.509604 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:58 crc kubenswrapper[4839]: I0227 19:49:58.510028 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:58 crc kubenswrapper[4839]: I0227 19:49:58.582483 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:59 crc kubenswrapper[4839]: I0227 19:49:59.169258 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:49:59 crc kubenswrapper[4839]: I0227 19:49:59.217693 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68nwh"] Feb 27 19:49:59 crc kubenswrapper[4839]: I0227 19:49:59.482709 4839 scope.go:117] "RemoveContainer" containerID="ff14ddc375cad7acab149b6cd2cedd080fbdeb204685691990524468298d4134" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.138283 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537030-t6rg8"] Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.139084 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.142016 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.142255 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.142338 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44lp8" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.154213 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537030-t6rg8"] Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.222496 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk4cn\" (UniqueName: \"kubernetes.io/projected/21377685-3db2-49e9-85b4-3e2c16b74bef-kube-api-access-xk4cn\") pod \"auto-csr-approver-29537030-t6rg8\" (UID: \"21377685-3db2-49e9-85b4-3e2c16b74bef\") " pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.323796 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk4cn\" (UniqueName: \"kubernetes.io/projected/21377685-3db2-49e9-85b4-3e2c16b74bef-kube-api-access-xk4cn\") pod \"auto-csr-approver-29537030-t6rg8\" (UID: \"21377685-3db2-49e9-85b4-3e2c16b74bef\") " pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.346536 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk4cn\" (UniqueName: \"kubernetes.io/projected/21377685-3db2-49e9-85b4-3e2c16b74bef-kube-api-access-xk4cn\") pod \"auto-csr-approver-29537030-t6rg8\" (UID: \"21377685-3db2-49e9-85b4-3e2c16b74bef\") " pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.501219 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:00 crc kubenswrapper[4839]: I0227 19:50:00.719915 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537030-t6rg8"] Feb 27 19:50:00 crc kubenswrapper[4839]: E0227 19:50:00.966078 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.134906 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" event={"ID":"21377685-3db2-49e9-85b4-3e2c16b74bef","Type":"ContainerStarted","Data":"e470ef154218b5b78d5da92dd7e9329f476b5525995b53483c835219ad9db659"} Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.135078 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-68nwh" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="registry-server" containerID="cri-o://84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03" gracePeriod=2 Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.547607 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.739825 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsb8d\" (UniqueName: \"kubernetes.io/projected/88a2d66b-54c2-401e-a822-1a166817b82e-kube-api-access-dsb8d\") pod \"88a2d66b-54c2-401e-a822-1a166817b82e\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.739897 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-catalog-content\") pod \"88a2d66b-54c2-401e-a822-1a166817b82e\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.740041 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-utilities\") pod \"88a2d66b-54c2-401e-a822-1a166817b82e\" (UID: \"88a2d66b-54c2-401e-a822-1a166817b82e\") " Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.741422 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-utilities" (OuterVolumeSpecName: "utilities") pod "88a2d66b-54c2-401e-a822-1a166817b82e" (UID: "88a2d66b-54c2-401e-a822-1a166817b82e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.749414 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a2d66b-54c2-401e-a822-1a166817b82e-kube-api-access-dsb8d" (OuterVolumeSpecName: "kube-api-access-dsb8d") pod "88a2d66b-54c2-401e-a822-1a166817b82e" (UID: "88a2d66b-54c2-401e-a822-1a166817b82e"). InnerVolumeSpecName "kube-api-access-dsb8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.841501 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:01 crc kubenswrapper[4839]: I0227 19:50:01.841532 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsb8d\" (UniqueName: \"kubernetes.io/projected/88a2d66b-54c2-401e-a822-1a166817b82e-kube-api-access-dsb8d\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.146111 4839 generic.go:334] "Generic (PLEG): container finished" podID="88a2d66b-54c2-401e-a822-1a166817b82e" containerID="84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03" exitCode=0 Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.146194 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerDied","Data":"84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03"} Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.146239 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68nwh" event={"ID":"88a2d66b-54c2-401e-a822-1a166817b82e","Type":"ContainerDied","Data":"c758200e6946203934c30750aa68c29134640f0593a4d03e31b9d43b1073826b"} Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.146297 4839 scope.go:117] "RemoveContainer" containerID="84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.146590 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68nwh" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.177615 4839 scope.go:117] "RemoveContainer" containerID="5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.209886 4839 scope.go:117] "RemoveContainer" containerID="ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.236404 4839 scope.go:117] "RemoveContainer" containerID="84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03" Feb 27 19:50:02 crc kubenswrapper[4839]: E0227 19:50:02.237065 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03\": container with ID starting with 84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03 not found: ID does not exist" containerID="84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.237117 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03"} err="failed to get container status \"84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03\": rpc error: code = NotFound desc = could not find container \"84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03\": container with ID starting with 84f926f0cf4f5212a87f562a74e1aceea05963a9867e42d9b6a757b5d7b28d03 not found: ID does not exist" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.237152 4839 scope.go:117] "RemoveContainer" containerID="5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203" Feb 27 19:50:02 crc kubenswrapper[4839]: E0227 19:50:02.237813 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203\": container with ID starting with 5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203 not found: ID does not exist" containerID="5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.237877 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203"} err="failed to get container status \"5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203\": rpc error: code = NotFound desc = could not find container \"5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203\": container with ID starting with 5b92f2c6e1dcff7f3172eb10f14250b153ab9cd1c12c4dd8eecbf01569491203 not found: ID does not exist" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.237914 4839 scope.go:117] "RemoveContainer" containerID="ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65" Feb 27 19:50:02 crc kubenswrapper[4839]: E0227 19:50:02.238433 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65\": container with ID starting with ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65 not found: ID does not exist" containerID="ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.238471 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65"} err="failed to get container status \"ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65\": rpc error: code = NotFound desc = could not find container \"ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65\": container with ID starting with ffeb2c17b2a757d424f2c3f52eba6badfb1b5c974245508f44dab90c867a4b65 not found: ID does not exist" Feb 27 19:50:02 crc kubenswrapper[4839]: E0227 19:50:02.341516 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:50:02 crc kubenswrapper[4839]: E0227 19:50:02.341731 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:50:02 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:50:02 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xk4cn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537030-t6rg8_openshift-infra(21377685-3db2-49e9-85b4-3e2c16b74bef): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:50:02 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:50:02 crc kubenswrapper[4839]: E0227 19:50:02.343297 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.794386 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88a2d66b-54c2-401e-a822-1a166817b82e" (UID: "88a2d66b-54c2-401e-a822-1a166817b82e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:50:02 crc kubenswrapper[4839]: I0227 19:50:02.863801 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a2d66b-54c2-401e-a822-1a166817b82e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:03 crc kubenswrapper[4839]: I0227 19:50:03.062133 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68nwh"] Feb 27 19:50:03 crc kubenswrapper[4839]: I0227 19:50:03.069197 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-68nwh"] Feb 27 19:50:03 crc kubenswrapper[4839]: E0227 19:50:03.154551 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" Feb 27 19:50:03 crc kubenswrapper[4839]: E0227 19:50:03.967793 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:50:04 crc kubenswrapper[4839]: I0227 19:50:04.972579 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" path="/var/lib/kubelet/pods/88a2d66b-54c2-401e-a822-1a166817b82e/volumes" Feb 27 19:50:09 crc kubenswrapper[4839]: E0227 19:50:09.966602 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:50:10 crc kubenswrapper[4839]: I0227 19:50:10.199649 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerStarted","Data":"61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd"} Feb 27 19:50:11 crc kubenswrapper[4839]: I0227 19:50:11.210657 4839 generic.go:334] "Generic (PLEG): container finished" podID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerID="61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd" exitCode=0 Feb 27 19:50:11 crc kubenswrapper[4839]: I0227 19:50:11.210736 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerDied","Data":"61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd"} Feb 27 19:50:12 crc kubenswrapper[4839]: I0227 19:50:12.220886 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerStarted","Data":"4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8"} Feb 27 19:50:12 crc kubenswrapper[4839]: I0227 19:50:12.247393 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9j7z5" podStartSLOduration=2.575193458 podStartE2EDuration="45.247373194s" podCreationTimestamp="2026-02-27 19:49:27 +0000 UTC" firstStartedPulling="2026-02-27 19:49:28.918699105 +0000 UTC m=+930.563568880" lastFinishedPulling="2026-02-27 19:50:11.590878841 +0000 UTC m=+973.235748616" observedRunningTime="2026-02-27 19:50:12.245895575 +0000 UTC m=+973.890765310" watchObservedRunningTime="2026-02-27 19:50:12.247373194 +0000 UTC m=+973.892242929" Feb 27 19:50:14 crc kubenswrapper[4839]: E0227 19:50:14.967893 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:50:17 crc kubenswrapper[4839]: I0227 19:50:17.397032 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:50:17 crc kubenswrapper[4839]: I0227 19:50:17.397113 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:50:17 crc kubenswrapper[4839]: E0227 19:50:17.880845 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:50:17 crc kubenswrapper[4839]: E0227 19:50:17.881422 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:50:17 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:50:17 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xk4cn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537030-t6rg8_openshift-infra(21377685-3db2-49e9-85b4-3e2c16b74bef): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:50:17 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:50:17 crc kubenswrapper[4839]: E0227 19:50:17.883116 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" Feb 27 19:50:18 crc kubenswrapper[4839]: I0227 19:50:18.445985 4839 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9j7z5" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="registry-server" probeResult="failure" output=< Feb 27 19:50:18 crc kubenswrapper[4839]: timeout: failed to connect service ":50051" within 1s Feb 27 19:50:18 crc kubenswrapper[4839]: > Feb 27 19:50:18 crc kubenswrapper[4839]: E0227 19:50:18.972216 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.955414 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v88w8"] Feb 27 19:50:21 crc kubenswrapper[4839]: E0227 19:50:21.956126 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="extract-content" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.956143 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="extract-content" Feb 27 19:50:21 crc kubenswrapper[4839]: E0227 19:50:21.956163 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="registry-server" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.956172 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="registry-server" Feb 27 19:50:21 crc kubenswrapper[4839]: E0227 19:50:21.956188 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="extract-utilities" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.956197 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="extract-utilities" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.956331 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a2d66b-54c2-401e-a822-1a166817b82e" containerName="registry-server" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.957358 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:21 crc kubenswrapper[4839]: I0227 19:50:21.965959 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v88w8"] Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.157245 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lc7l\" (UniqueName: \"kubernetes.io/projected/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-kube-api-access-2lc7l\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.157411 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-utilities\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.157432 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-catalog-content\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.258488 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-utilities\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.258541 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-catalog-content\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.258587 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lc7l\" (UniqueName: \"kubernetes.io/projected/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-kube-api-access-2lc7l\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.259092 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-utilities\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.259104 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-catalog-content\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.282051 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lc7l\" (UniqueName: \"kubernetes.io/projected/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-kube-api-access-2lc7l\") pod \"community-operators-v88w8\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.579077 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.682617 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.682745 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:50:22 crc kubenswrapper[4839]: I0227 19:50:22.820217 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v88w8"] Feb 27 19:50:22 crc kubenswrapper[4839]: E0227 19:50:22.966893 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:50:23 crc kubenswrapper[4839]: I0227 19:50:23.289249 4839 generic.go:334] "Generic (PLEG): container finished" podID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerID="443e0ef2d86653eac9b1729f2d09a218b8b5cd58146a04237a794c9794f26e0b" exitCode=0 Feb 27 19:50:23 crc kubenswrapper[4839]: I0227 19:50:23.289307 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerDied","Data":"443e0ef2d86653eac9b1729f2d09a218b8b5cd58146a04237a794c9794f26e0b"} Feb 27 19:50:23 crc kubenswrapper[4839]: I0227 19:50:23.289344 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerStarted","Data":"ddabfb891c67dc22305081fbf8df023dc26911ec91eaf7541bc7b4ef28664065"} Feb 27 19:50:24 crc kubenswrapper[4839]: I0227 19:50:24.296643 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerStarted","Data":"78bf4c35fb6f7fa7b8e6963ae165ecb19189c0f22867e88d5d3f9b950d319211"} Feb 27 19:50:25 crc kubenswrapper[4839]: I0227 19:50:25.306824 4839 generic.go:334] "Generic (PLEG): container finished" podID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerID="78bf4c35fb6f7fa7b8e6963ae165ecb19189c0f22867e88d5d3f9b950d319211" exitCode=0 Feb 27 19:50:25 crc kubenswrapper[4839]: I0227 19:50:25.306879 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerDied","Data":"78bf4c35fb6f7fa7b8e6963ae165ecb19189c0f22867e88d5d3f9b950d319211"} Feb 27 19:50:25 crc kubenswrapper[4839]: E0227 19:50:25.967080 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:50:26 crc kubenswrapper[4839]: I0227 19:50:26.315173 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerStarted","Data":"e14f28d8a349567a1c7aab696c7720c73a6bd20667d0c57927142627b208afe3"} Feb 27 19:50:26 crc kubenswrapper[4839]: I0227 19:50:26.334561 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v88w8" podStartSLOduration=2.520325492 podStartE2EDuration="5.334540885s" podCreationTimestamp="2026-02-27 19:50:21 +0000 UTC" firstStartedPulling="2026-02-27 19:50:23.290555058 +0000 UTC m=+984.935424793" lastFinishedPulling="2026-02-27 19:50:26.104770421 +0000 UTC m=+987.749640186" observedRunningTime="2026-02-27 19:50:26.33434135 +0000 UTC m=+987.979211115" watchObservedRunningTime="2026-02-27 19:50:26.334540885 +0000 UTC m=+987.979410620" Feb 27 19:50:27 crc kubenswrapper[4839]: I0227 19:50:27.457582 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:50:27 crc kubenswrapper[4839]: I0227 19:50:27.513579 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.315129 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9j7z5"] Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.331553 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9j7z5" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="registry-server" containerID="cri-o://4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8" gracePeriod=2 Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.702118 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.765558 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-catalog-content\") pod \"b64b6367-e435-4dac-aa22-ac7efafdcd21\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.765626 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-utilities\") pod \"b64b6367-e435-4dac-aa22-ac7efafdcd21\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.765691 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s7kt\" (UniqueName: \"kubernetes.io/projected/b64b6367-e435-4dac-aa22-ac7efafdcd21-kube-api-access-5s7kt\") pod \"b64b6367-e435-4dac-aa22-ac7efafdcd21\" (UID: \"b64b6367-e435-4dac-aa22-ac7efafdcd21\") " Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.766469 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-utilities" (OuterVolumeSpecName: "utilities") pod "b64b6367-e435-4dac-aa22-ac7efafdcd21" (UID: "b64b6367-e435-4dac-aa22-ac7efafdcd21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.774080 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b64b6367-e435-4dac-aa22-ac7efafdcd21-kube-api-access-5s7kt" (OuterVolumeSpecName: "kube-api-access-5s7kt") pod "b64b6367-e435-4dac-aa22-ac7efafdcd21" (UID: "b64b6367-e435-4dac-aa22-ac7efafdcd21"). InnerVolumeSpecName "kube-api-access-5s7kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.867512 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.867582 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s7kt\" (UniqueName: \"kubernetes.io/projected/b64b6367-e435-4dac-aa22-ac7efafdcd21-kube-api-access-5s7kt\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.890168 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b64b6367-e435-4dac-aa22-ac7efafdcd21" (UID: "b64b6367-e435-4dac-aa22-ac7efafdcd21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:50:29 crc kubenswrapper[4839]: I0227 19:50:29.969111 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64b6367-e435-4dac-aa22-ac7efafdcd21-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.340957 4839 generic.go:334] "Generic (PLEG): container finished" podID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerID="4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8" exitCode=0 Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.341054 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerDied","Data":"4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8"} Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.341109 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9j7z5" event={"ID":"b64b6367-e435-4dac-aa22-ac7efafdcd21","Type":"ContainerDied","Data":"83e3b6ceb67a93c006aa863a34483d4e75d5de16fc01d14d24d1cf2d8b00216b"} Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.341143 4839 scope.go:117] "RemoveContainer" containerID="4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.341398 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9j7z5" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.376777 4839 scope.go:117] "RemoveContainer" containerID="61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.390814 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9j7z5"] Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.406361 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9j7z5"] Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.407653 4839 scope.go:117] "RemoveContainer" containerID="86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.430128 4839 scope.go:117] "RemoveContainer" containerID="4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8" Feb 27 19:50:30 crc kubenswrapper[4839]: E0227 19:50:30.430903 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8\": container with ID starting with 4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8 not found: ID does not exist" containerID="4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.430985 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8"} err="failed to get container status \"4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8\": rpc error: code = NotFound desc = could not find container \"4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8\": container with ID starting with 4027d00301ec433fdd861c0d7a2081dc47e6f87640060f236946eb58644b0af8 not found: ID does not exist" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.431043 4839 scope.go:117] "RemoveContainer" containerID="61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd" Feb 27 19:50:30 crc kubenswrapper[4839]: E0227 19:50:30.431840 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd\": container with ID starting with 61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd not found: ID does not exist" containerID="61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.431902 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd"} err="failed to get container status \"61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd\": rpc error: code = NotFound desc = could not find container \"61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd\": container with ID starting with 61601beaed30ffb1bf2af4e68d563b18e7586d650f491b7073f32e69525fcffd not found: ID does not exist" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.431945 4839 scope.go:117] "RemoveContainer" containerID="86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127" Feb 27 19:50:30 crc kubenswrapper[4839]: E0227 19:50:30.432503 4839 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127\": container with ID starting with 86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127 not found: ID does not exist" containerID="86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.432615 4839 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127"} err="failed to get container status \"86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127\": rpc error: code = NotFound desc = could not find container \"86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127\": container with ID starting with 86fccb6ad3ec6858746f2a7daad4f779636691cd7b06ce0ddfb39406c821c127 not found: ID does not exist" Feb 27 19:50:30 crc kubenswrapper[4839]: E0227 19:50:30.967965 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" Feb 27 19:50:30 crc kubenswrapper[4839]: I0227 19:50:30.971789 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" path="/var/lib/kubelet/pods/b64b6367-e435-4dac-aa22-ac7efafdcd21/volumes" Feb 27 19:50:32 crc kubenswrapper[4839]: I0227 19:50:32.579381 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:32 crc kubenswrapper[4839]: I0227 19:50:32.579779 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:32 crc kubenswrapper[4839]: I0227 19:50:32.628287 4839 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:32 crc kubenswrapper[4839]: E0227 19:50:32.967455 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:50:33 crc kubenswrapper[4839]: I0227 19:50:33.415939 4839 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:34 crc kubenswrapper[4839]: I0227 19:50:34.519890 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v88w8"] Feb 27 19:50:35 crc kubenswrapper[4839]: I0227 19:50:35.376644 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v88w8" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="registry-server" containerID="cri-o://e14f28d8a349567a1c7aab696c7720c73a6bd20667d0c57927142627b208afe3" gracePeriod=2 Feb 27 19:50:35 crc kubenswrapper[4839]: E0227 19:50:35.967347 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.396598 4839 generic.go:334] "Generic (PLEG): container finished" podID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerID="e14f28d8a349567a1c7aab696c7720c73a6bd20667d0c57927142627b208afe3" exitCode=0 Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.396616 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerDied","Data":"e14f28d8a349567a1c7aab696c7720c73a6bd20667d0c57927142627b208afe3"} Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.882438 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.964707 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lc7l\" (UniqueName: \"kubernetes.io/projected/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-kube-api-access-2lc7l\") pod \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.964781 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-catalog-content\") pod \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.964855 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-utilities\") pod \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\" (UID: \"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3\") " Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.966015 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-utilities" (OuterVolumeSpecName: "utilities") pod "01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" (UID: "01920ee3-85b3-4a5b-bd6a-5a655a1af7b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:50:36 crc kubenswrapper[4839]: I0227 19:50:36.974017 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-kube-api-access-2lc7l" (OuterVolumeSpecName: "kube-api-access-2lc7l") pod "01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" (UID: "01920ee3-85b3-4a5b-bd6a-5a655a1af7b3"). InnerVolumeSpecName "kube-api-access-2lc7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.017650 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" (UID: "01920ee3-85b3-4a5b-bd6a-5a655a1af7b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.065679 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lc7l\" (UniqueName: \"kubernetes.io/projected/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-kube-api-access-2lc7l\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.065709 4839 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.065719 4839 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.404573 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v88w8" event={"ID":"01920ee3-85b3-4a5b-bd6a-5a655a1af7b3","Type":"ContainerDied","Data":"ddabfb891c67dc22305081fbf8df023dc26911ec91eaf7541bc7b4ef28664065"} Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.404638 4839 scope.go:117] "RemoveContainer" containerID="e14f28d8a349567a1c7aab696c7720c73a6bd20667d0c57927142627b208afe3" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.404852 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v88w8" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.436202 4839 scope.go:117] "RemoveContainer" containerID="78bf4c35fb6f7fa7b8e6963ae165ecb19189c0f22867e88d5d3f9b950d319211" Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.444021 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v88w8"] Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.447711 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v88w8"] Feb 27 19:50:37 crc kubenswrapper[4839]: I0227 19:50:37.464036 4839 scope.go:117] "RemoveContainer" containerID="443e0ef2d86653eac9b1729f2d09a218b8b5cd58146a04237a794c9794f26e0b" Feb 27 19:50:38 crc kubenswrapper[4839]: I0227 19:50:38.977324 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" path="/var/lib/kubelet/pods/01920ee3-85b3-4a5b-bd6a-5a655a1af7b3/volumes" Feb 27 19:50:40 crc kubenswrapper[4839]: E0227 19:50:40.536396 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:50:40 crc kubenswrapper[4839]: E0227 19:50:40.536540 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzh9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7hm7c_openshift-marketplace(02ff6284-1eb0-403a-8bfd-67184d2b3f5e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:50:40 crc kubenswrapper[4839]: E0227 19:50:40.537737 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:50:44 crc kubenswrapper[4839]: I0227 19:50:44.450925 4839 generic.go:334] "Generic (PLEG): container finished" podID="21377685-3db2-49e9-85b4-3e2c16b74bef" containerID="2f0c07374977a5f9b6b5fb42df471d67c7922c6a68de76151426573c6cd2f8f8" exitCode=0 Feb 27 19:50:44 crc kubenswrapper[4839]: I0227 19:50:44.450970 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" event={"ID":"21377685-3db2-49e9-85b4-3e2c16b74bef","Type":"ContainerDied","Data":"2f0c07374977a5f9b6b5fb42df471d67c7922c6a68de76151426573c6cd2f8f8"} Feb 27 19:50:45 crc kubenswrapper[4839]: I0227 19:50:45.734880 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:45 crc kubenswrapper[4839]: I0227 19:50:45.888013 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk4cn\" (UniqueName: \"kubernetes.io/projected/21377685-3db2-49e9-85b4-3e2c16b74bef-kube-api-access-xk4cn\") pod \"21377685-3db2-49e9-85b4-3e2c16b74bef\" (UID: \"21377685-3db2-49e9-85b4-3e2c16b74bef\") " Feb 27 19:50:45 crc kubenswrapper[4839]: I0227 19:50:45.894690 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21377685-3db2-49e9-85b4-3e2c16b74bef-kube-api-access-xk4cn" (OuterVolumeSpecName: "kube-api-access-xk4cn") pod "21377685-3db2-49e9-85b4-3e2c16b74bef" (UID: "21377685-3db2-49e9-85b4-3e2c16b74bef"). InnerVolumeSpecName "kube-api-access-xk4cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:50:45 crc kubenswrapper[4839]: E0227 19:50:45.967233 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:50:45 crc kubenswrapper[4839]: I0227 19:50:45.989534 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk4cn\" (UniqueName: \"kubernetes.io/projected/21377685-3db2-49e9-85b4-3e2c16b74bef-kube-api-access-xk4cn\") on node \"crc\" DevicePath \"\"" Feb 27 19:50:46 crc kubenswrapper[4839]: I0227 19:50:46.468564 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" event={"ID":"21377685-3db2-49e9-85b4-3e2c16b74bef","Type":"ContainerDied","Data":"e470ef154218b5b78d5da92dd7e9329f476b5525995b53483c835219ad9db659"} Feb 27 19:50:46 crc kubenswrapper[4839]: I0227 19:50:46.468627 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e470ef154218b5b78d5da92dd7e9329f476b5525995b53483c835219ad9db659" Feb 27 19:50:46 crc kubenswrapper[4839]: I0227 19:50:46.468655 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537030-t6rg8" Feb 27 19:50:46 crc kubenswrapper[4839]: I0227 19:50:46.803887 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537024-szhk2"] Feb 27 19:50:46 crc kubenswrapper[4839]: I0227 19:50:46.811926 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537024-szhk2"] Feb 27 19:50:46 crc kubenswrapper[4839]: I0227 19:50:46.975375 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45766657-34fd-4b18-a32b-8e71a1c711e2" path="/var/lib/kubelet/pods/45766657-34fd-4b18-a32b-8e71a1c711e2/volumes" Feb 27 19:50:49 crc kubenswrapper[4839]: E0227 19:50:49.967643 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:50:52 crc kubenswrapper[4839]: I0227 19:50:52.682346 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:50:52 crc kubenswrapper[4839]: I0227 19:50:52.682690 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:50:52 crc kubenswrapper[4839]: I0227 19:50:52.682734 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:50:52 crc kubenswrapper[4839]: I0227 19:50:52.683163 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cebc994c5b008539c690e93a6330de8ce7ecd9ea0602f50a6d17fdfe769f11e8"} pod="openshift-machine-config-operator/machine-config-daemon-thb8n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 19:50:52 crc kubenswrapper[4839]: I0227 19:50:52.683213 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" containerID="cri-o://cebc994c5b008539c690e93a6330de8ce7ecd9ea0602f50a6d17fdfe769f11e8" gracePeriod=600 Feb 27 19:50:53 crc kubenswrapper[4839]: I0227 19:50:53.521546 4839 generic.go:334] "Generic (PLEG): container finished" podID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerID="cebc994c5b008539c690e93a6330de8ce7ecd9ea0602f50a6d17fdfe769f11e8" exitCode=0 Feb 27 19:50:53 crc kubenswrapper[4839]: I0227 19:50:53.521626 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerDied","Data":"cebc994c5b008539c690e93a6330de8ce7ecd9ea0602f50a6d17fdfe769f11e8"} Feb 27 19:50:53 crc kubenswrapper[4839]: I0227 19:50:53.521878 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"324b7c3677fa3a46e6e785649ed814cb180ff19bad6941d855be7fc5a76021ea"} Feb 27 19:50:53 crc kubenswrapper[4839]: I0227 19:50:53.521900 4839 scope.go:117] "RemoveContainer" containerID="5fc60b748d670020094881ec089d53e0f67db6413e9b475783fdc7432a738c18" Feb 27 19:50:54 crc kubenswrapper[4839]: E0227 19:50:54.968159 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:50:58 crc kubenswrapper[4839]: E0227 19:50:58.978140 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:50:59 crc kubenswrapper[4839]: I0227 19:50:59.545386 4839 scope.go:117] "RemoveContainer" containerID="fd4d7d181beb069c7d7c147ddeb49b7444289e329deb312d5a53757520498513" Feb 27 19:51:02 crc kubenswrapper[4839]: E0227 19:51:02.967491 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.845807 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n8qcq/must-gather-2cl2j"] Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846118 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" containerName="oc" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846148 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" containerName="oc" Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846169 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="registry-server" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846184 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="registry-server" Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846203 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="extract-utilities" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846216 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="extract-utilities" Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846235 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="extract-content" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846246 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="extract-content" Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846269 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="registry-server" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846282 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="registry-server" Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846299 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="extract-content" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846311 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="extract-content" Feb 27 19:51:03 crc kubenswrapper[4839]: E0227 19:51:03.846330 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="extract-utilities" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846344 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="extract-utilities" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846507 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64b6367-e435-4dac-aa22-ac7efafdcd21" containerName="registry-server" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846524 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="21377685-3db2-49e9-85b4-3e2c16b74bef" containerName="oc" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.846540 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="01920ee3-85b3-4a5b-bd6a-5a655a1af7b3" containerName="registry-server" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.847488 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.849121 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n8qcq"/"openshift-service-ca.crt" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.849322 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n8qcq"/"kube-root-ca.crt" Feb 27 19:51:03 crc kubenswrapper[4839]: I0227 19:51:03.860929 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n8qcq/must-gather-2cl2j"] Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.028530 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5f1721c-699e-4d94-98aa-637641275f8b-must-gather-output\") pod \"must-gather-2cl2j\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.028632 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbgqb\" (UniqueName: \"kubernetes.io/projected/a5f1721c-699e-4d94-98aa-637641275f8b-kube-api-access-qbgqb\") pod \"must-gather-2cl2j\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.129505 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5f1721c-699e-4d94-98aa-637641275f8b-must-gather-output\") pod \"must-gather-2cl2j\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.129578 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbgqb\" (UniqueName: \"kubernetes.io/projected/a5f1721c-699e-4d94-98aa-637641275f8b-kube-api-access-qbgqb\") pod \"must-gather-2cl2j\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.129928 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5f1721c-699e-4d94-98aa-637641275f8b-must-gather-output\") pod \"must-gather-2cl2j\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.145578 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbgqb\" (UniqueName: \"kubernetes.io/projected/a5f1721c-699e-4d94-98aa-637641275f8b-kube-api-access-qbgqb\") pod \"must-gather-2cl2j\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.172440 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.394988 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n8qcq/must-gather-2cl2j"] Feb 27 19:51:04 crc kubenswrapper[4839]: I0227 19:51:04.603650 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" event={"ID":"a5f1721c-699e-4d94-98aa-637641275f8b","Type":"ContainerStarted","Data":"84024a785aaf9b2c781f8263cf201c680ee6324a3bb85cef80898781499ae3df"} Feb 27 19:51:08 crc kubenswrapper[4839]: E0227 19:51:08.979206 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:51:09 crc kubenswrapper[4839]: I0227 19:51:09.633725 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" event={"ID":"a5f1721c-699e-4d94-98aa-637641275f8b","Type":"ContainerStarted","Data":"ba83ad029116bd8126800a43ff2cdda4ea92c90d2cd7322221dba58de9d58392"} Feb 27 19:51:09 crc kubenswrapper[4839]: I0227 19:51:09.634201 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" event={"ID":"a5f1721c-699e-4d94-98aa-637641275f8b","Type":"ContainerStarted","Data":"59698af33f6df4a7be343651e34c975446cf507569fe8c9bad55ec5d702f3e88"} Feb 27 19:51:09 crc kubenswrapper[4839]: I0227 19:51:09.648931 4839 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" podStartSLOduration=2.159131922 podStartE2EDuration="6.648905437s" podCreationTimestamp="2026-02-27 19:51:03 +0000 UTC" firstStartedPulling="2026-02-27 19:51:04.404941798 +0000 UTC m=+1026.049811533" lastFinishedPulling="2026-02-27 19:51:08.894715303 +0000 UTC m=+1030.539585048" observedRunningTime="2026-02-27 19:51:09.647072768 +0000 UTC m=+1031.291942513" watchObservedRunningTime="2026-02-27 19:51:09.648905437 +0000 UTC m=+1031.293775172" Feb 27 19:51:09 crc kubenswrapper[4839]: E0227 19:51:09.967267 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:51:15 crc kubenswrapper[4839]: E0227 19:51:15.967877 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:51:19 crc kubenswrapper[4839]: E0227 19:51:19.971054 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:51:20 crc kubenswrapper[4839]: E0227 19:51:20.967182 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:51:27 crc kubenswrapper[4839]: E0227 19:51:27.968131 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:51:33 crc kubenswrapper[4839]: E0227 19:51:33.966646 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:51:34 crc kubenswrapper[4839]: E0227 19:51:34.966854 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:51:41 crc kubenswrapper[4839]: E0227 19:51:41.966997 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:51:45 crc kubenswrapper[4839]: I0227 19:51:45.746176 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-kncvh_d51493b0-bd22-4a0e-9f54-c8a980289a4d/control-plane-machine-set-operator/0.log" Feb 27 19:51:45 crc kubenswrapper[4839]: I0227 19:51:45.878624 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pgbmd_22e4cd58-dab8-41fa-9cae-323ce54a4b0d/kube-rbac-proxy/0.log" Feb 27 19:51:45 crc kubenswrapper[4839]: I0227 19:51:45.897333 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pgbmd_22e4cd58-dab8-41fa-9cae-323ce54a4b0d/machine-api-operator/0.log" Feb 27 19:51:47 crc kubenswrapper[4839]: E0227 19:51:47.968332 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:51:48 crc kubenswrapper[4839]: E0227 19:51:48.969380 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:51:54 crc kubenswrapper[4839]: E0227 19:51:54.967411 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:51:57 crc kubenswrapper[4839]: I0227 19:51:57.884580 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-pd6vc_9c95be49-7215-4016-8d89-0dd008c7ba2f/cert-manager-controller/0.log" Feb 27 19:51:58 crc kubenswrapper[4839]: I0227 19:51:58.021749 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-7rchl_58e0aaec-ea1f-46f9-a599-7ead3e24341d/cert-manager-cainjector/0.log" Feb 27 19:51:58 crc kubenswrapper[4839]: I0227 19:51:58.045032 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-gxszm_56781d07-f0fd-4d2e-97d4-784c15af5737/cert-manager-webhook/0.log" Feb 27 19:51:59 crc kubenswrapper[4839]: I0227 19:51:59.613924 4839 scope.go:117] "RemoveContainer" containerID="bb374ab362b865917efe6077abab0be2ffca71aac8403dd4e8db06cda9145b4a" Feb 27 19:51:59 crc kubenswrapper[4839]: E0227 19:51:59.966132 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.133302 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537032-wvf9m"] Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.134177 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.135971 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.136186 4839 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44lp8" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.136358 4839 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.144131 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537032-wvf9m"] Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.235577 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmf4r\" (UniqueName: \"kubernetes.io/projected/4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97-kube-api-access-hmf4r\") pod \"auto-csr-approver-29537032-wvf9m\" (UID: \"4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97\") " pod="openshift-infra/auto-csr-approver-29537032-wvf9m" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.336901 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmf4r\" (UniqueName: \"kubernetes.io/projected/4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97-kube-api-access-hmf4r\") pod \"auto-csr-approver-29537032-wvf9m\" (UID: \"4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97\") " pod="openshift-infra/auto-csr-approver-29537032-wvf9m" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.355136 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmf4r\" (UniqueName: \"kubernetes.io/projected/4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97-kube-api-access-hmf4r\") pod \"auto-csr-approver-29537032-wvf9m\" (UID: \"4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97\") " pod="openshift-infra/auto-csr-approver-29537032-wvf9m" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.452552 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.856131 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537032-wvf9m"] Feb 27 19:52:00 crc kubenswrapper[4839]: I0227 19:52:00.928089 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" event={"ID":"4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97","Type":"ContainerStarted","Data":"c6720a60192fcb1d59a5fa9a268bde8d13ed975d72754c530208b9dd1d360efb"} Feb 27 19:52:01 crc kubenswrapper[4839]: E0227 19:52:01.818484 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:52:01 crc kubenswrapper[4839]: E0227 19:52:01.819030 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:52:01 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:52:01 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmf4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537032-wvf9m_openshift-infra(4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:52:01 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:52:01 crc kubenswrapper[4839]: E0227 19:52:01.820629 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:52:01 crc kubenswrapper[4839]: E0227 19:52:01.934796 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:52:04 crc kubenswrapper[4839]: E0227 19:52:04.827530 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 19:52:04 crc kubenswrapper[4839]: E0227 19:52:04.827946 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l8rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vmj_openshift-marketplace(827f2124-61ed-4c5e-acf7-21463af06877): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:52:04 crc kubenswrapper[4839]: E0227 19:52:04.829139 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:52:09 crc kubenswrapper[4839]: E0227 19:52:09.608159 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:52:09 crc kubenswrapper[4839]: E0227 19:52:09.608706 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pfgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mngj4_openshift-marketplace(babba9f2-c281-4ad8-b697-9eb744a1960e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:52:09 crc kubenswrapper[4839]: E0227 19:52:09.609911 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:52:14 crc kubenswrapper[4839]: E0227 19:52:14.967317 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:52:16 crc kubenswrapper[4839]: E0227 19:52:16.969676 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:52:18 crc kubenswrapper[4839]: E0227 19:52:18.004165 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:52:18 crc kubenswrapper[4839]: E0227 19:52:18.004524 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:52:18 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:52:18 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmf4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537032-wvf9m_openshift-infra(4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:52:18 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:52:18 crc kubenswrapper[4839]: E0227 19:52:18.005722 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:52:20 crc kubenswrapper[4839]: E0227 19:52:20.966712 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:52:22 crc kubenswrapper[4839]: I0227 19:52:22.782630 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/extract-utilities/0.log" Feb 27 19:52:22 crc kubenswrapper[4839]: I0227 19:52:22.921305 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/extract-utilities/0.log" Feb 27 19:52:22 crc kubenswrapper[4839]: I0227 19:52:22.931986 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/extract-content/0.log" Feb 27 19:52:22 crc kubenswrapper[4839]: I0227 19:52:22.945113 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/extract-content/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.095438 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/extract-content/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.107010 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/extract-utilities/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.272235 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/extract-utilities/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.330990 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fqhwp_e640d992-0a39-4fd3-b53b-5a8afe9a5a4a/registry-server/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.405831 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/extract-utilities/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.414805 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/extract-content/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.438701 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/extract-content/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.587328 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/extract-content/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.593576 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/extract-utilities/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.740910 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gdwpn_ccee8c8e-aa9b-4aa4-8f0d-03e56a3c33c7/marketplace-operator/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.774811 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f5dwg_7986d862-dbf4-452a-a1e7-2d02cc850d76/registry-server/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.811003 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7hm7c_02ff6284-1eb0-403a-8bfd-67184d2b3f5e/extract-utilities/0.log" Feb 27 19:52:23 crc kubenswrapper[4839]: I0227 19:52:23.970962 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7hm7c_02ff6284-1eb0-403a-8bfd-67184d2b3f5e/extract-utilities/0.log" Feb 27 19:52:24 crc kubenswrapper[4839]: I0227 19:52:24.127090 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7hm7c_02ff6284-1eb0-403a-8bfd-67184d2b3f5e/extract-utilities/0.log" Feb 27 19:52:24 crc kubenswrapper[4839]: I0227 19:52:24.274297 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mngj4_babba9f2-c281-4ad8-b697-9eb744a1960e/extract-utilities/0.log" Feb 27 19:52:24 crc kubenswrapper[4839]: I0227 19:52:24.451899 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mngj4_babba9f2-c281-4ad8-b697-9eb744a1960e/extract-utilities/0.log" Feb 27 19:52:24 crc kubenswrapper[4839]: I0227 19:52:24.620079 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mngj4_babba9f2-c281-4ad8-b697-9eb744a1960e/extract-utilities/0.log" Feb 27 19:52:24 crc kubenswrapper[4839]: I0227 19:52:24.769038 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7vmj_827f2124-61ed-4c5e-acf7-21463af06877/extract-utilities/0.log" Feb 27 19:52:24 crc kubenswrapper[4839]: I0227 19:52:24.914393 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7vmj_827f2124-61ed-4c5e-acf7-21463af06877/extract-utilities/0.log" Feb 27 19:52:25 crc kubenswrapper[4839]: I0227 19:52:25.047009 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7vmj_827f2124-61ed-4c5e-acf7-21463af06877/extract-utilities/0.log" Feb 27 19:52:26 crc kubenswrapper[4839]: E0227 19:52:26.968217 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:52:28 crc kubenswrapper[4839]: E0227 19:52:28.973015 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:52:31 crc kubenswrapper[4839]: E0227 19:52:31.967275 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:52:34 crc kubenswrapper[4839]: E0227 19:52:34.967273 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:52:37 crc kubenswrapper[4839]: E0227 19:52:37.968186 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:52:39 crc kubenswrapper[4839]: E0227 19:52:39.967986 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:52:45 crc kubenswrapper[4839]: E0227 19:52:45.827372 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:52:45 crc kubenswrapper[4839]: E0227 19:52:45.828170 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:52:45 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:52:45 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmf4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537032-wvf9m_openshift-infra(4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:52:45 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:52:45 crc kubenswrapper[4839]: E0227 19:52:45.829948 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:52:47 crc kubenswrapper[4839]: E0227 19:52:47.968226 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:52:49 crc kubenswrapper[4839]: E0227 19:52:49.967985 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:52:52 crc kubenswrapper[4839]: I0227 19:52:52.681466 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:52:52 crc kubenswrapper[4839]: I0227 19:52:52.681806 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:52:52 crc kubenswrapper[4839]: E0227 19:52:52.969097 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:52:59 crc kubenswrapper[4839]: E0227 19:52:59.967791 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:53:02 crc kubenswrapper[4839]: E0227 19:53:02.967321 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:53:03 crc kubenswrapper[4839]: E0227 19:53:03.966315 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:53:07 crc kubenswrapper[4839]: E0227 19:53:07.967390 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:53:10 crc kubenswrapper[4839]: I0227 19:53:10.307076 4839 generic.go:334] "Generic (PLEG): container finished" podID="a5f1721c-699e-4d94-98aa-637641275f8b" containerID="59698af33f6df4a7be343651e34c975446cf507569fe8c9bad55ec5d702f3e88" exitCode=0 Feb 27 19:53:10 crc kubenswrapper[4839]: I0227 19:53:10.307143 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" event={"ID":"a5f1721c-699e-4d94-98aa-637641275f8b","Type":"ContainerDied","Data":"59698af33f6df4a7be343651e34c975446cf507569fe8c9bad55ec5d702f3e88"} Feb 27 19:53:10 crc kubenswrapper[4839]: I0227 19:53:10.307784 4839 scope.go:117] "RemoveContainer" containerID="59698af33f6df4a7be343651e34c975446cf507569fe8c9bad55ec5d702f3e88" Feb 27 19:53:10 crc kubenswrapper[4839]: I0227 19:53:10.803505 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n8qcq_must-gather-2cl2j_a5f1721c-699e-4d94-98aa-637641275f8b/gather/0.log" Feb 27 19:53:11 crc kubenswrapper[4839]: E0227 19:53:11.966377 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:53:13 crc kubenswrapper[4839]: E0227 19:53:13.968231 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:53:16 crc kubenswrapper[4839]: E0227 19:53:16.968182 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.037488 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n8qcq/must-gather-2cl2j"] Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.037877 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="copy" containerID="cri-o://ba83ad029116bd8126800a43ff2cdda4ea92c90d2cd7322221dba58de9d58392" gracePeriod=2 Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.043064 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n8qcq/must-gather-2cl2j"] Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.360863 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n8qcq_must-gather-2cl2j_a5f1721c-699e-4d94-98aa-637641275f8b/copy/0.log" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.362287 4839 generic.go:334] "Generic (PLEG): container finished" podID="a5f1721c-699e-4d94-98aa-637641275f8b" containerID="ba83ad029116bd8126800a43ff2cdda4ea92c90d2cd7322221dba58de9d58392" exitCode=143 Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.407902 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n8qcq_must-gather-2cl2j_a5f1721c-699e-4d94-98aa-637641275f8b/copy/0.log" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.408367 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.516316 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbgqb\" (UniqueName: \"kubernetes.io/projected/a5f1721c-699e-4d94-98aa-637641275f8b-kube-api-access-qbgqb\") pod \"a5f1721c-699e-4d94-98aa-637641275f8b\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.516423 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5f1721c-699e-4d94-98aa-637641275f8b-must-gather-output\") pod \"a5f1721c-699e-4d94-98aa-637641275f8b\" (UID: \"a5f1721c-699e-4d94-98aa-637641275f8b\") " Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.531866 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f1721c-699e-4d94-98aa-637641275f8b-kube-api-access-qbgqb" (OuterVolumeSpecName: "kube-api-access-qbgqb") pod "a5f1721c-699e-4d94-98aa-637641275f8b" (UID: "a5f1721c-699e-4d94-98aa-637641275f8b"). InnerVolumeSpecName "kube-api-access-qbgqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.564034 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5f1721c-699e-4d94-98aa-637641275f8b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a5f1721c-699e-4d94-98aa-637641275f8b" (UID: "a5f1721c-699e-4d94-98aa-637641275f8b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.617437 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbgqb\" (UniqueName: \"kubernetes.io/projected/a5f1721c-699e-4d94-98aa-637641275f8b-kube-api-access-qbgqb\") on node \"crc\" DevicePath \"\"" Feb 27 19:53:17 crc kubenswrapper[4839]: I0227 19:53:17.617467 4839 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5f1721c-699e-4d94-98aa-637641275f8b-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 27 19:53:18 crc kubenswrapper[4839]: I0227 19:53:18.369747 4839 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n8qcq_must-gather-2cl2j_a5f1721c-699e-4d94-98aa-637641275f8b/copy/0.log" Feb 27 19:53:18 crc kubenswrapper[4839]: I0227 19:53:18.370987 4839 scope.go:117] "RemoveContainer" containerID="ba83ad029116bd8126800a43ff2cdda4ea92c90d2cd7322221dba58de9d58392" Feb 27 19:53:18 crc kubenswrapper[4839]: I0227 19:53:18.371022 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n8qcq/must-gather-2cl2j" Feb 27 19:53:18 crc kubenswrapper[4839]: I0227 19:53:18.387610 4839 scope.go:117] "RemoveContainer" containerID="59698af33f6df4a7be343651e34c975446cf507569fe8c9bad55ec5d702f3e88" Feb 27 19:53:18 crc kubenswrapper[4839]: I0227 19:53:18.982938 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" path="/var/lib/kubelet/pods/a5f1721c-699e-4d94-98aa-637641275f8b/volumes" Feb 27 19:53:22 crc kubenswrapper[4839]: I0227 19:53:22.682422 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:53:22 crc kubenswrapper[4839]: I0227 19:53:22.682822 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:53:22 crc kubenswrapper[4839]: E0227 19:53:22.969118 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:53:25 crc kubenswrapper[4839]: I0227 19:53:25.968274 4839 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 19:53:26 crc kubenswrapper[4839]: E0227 19:53:26.943324 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:53:26 crc kubenswrapper[4839]: E0227 19:53:26.943493 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:53:26 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:53:26 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmf4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537032-wvf9m_openshift-infra(4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:53:26 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:53:26 crc kubenswrapper[4839]: E0227 19:53:26.944636 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:53:27 crc kubenswrapper[4839]: E0227 19:53:27.968076 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:53:31 crc kubenswrapper[4839]: E0227 19:53:31.650204 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 19:53:31 crc kubenswrapper[4839]: E0227 19:53:31.650444 4839 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzh9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7hm7c_openshift-marketplace(02ff6284-1eb0-403a-8bfd-67184d2b3f5e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 19:53:31 crc kubenswrapper[4839]: E0227 19:53:31.651785 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:53:33 crc kubenswrapper[4839]: E0227 19:53:33.967250 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:53:37 crc kubenswrapper[4839]: E0227 19:53:37.969364 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:53:40 crc kubenswrapper[4839]: E0227 19:53:40.970758 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:53:42 crc kubenswrapper[4839]: E0227 19:53:42.967613 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:53:44 crc kubenswrapper[4839]: E0227 19:53:44.966175 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:53:51 crc kubenswrapper[4839]: E0227 19:53:51.966488 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:53:52 crc kubenswrapper[4839]: I0227 19:53:52.682146 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:53:52 crc kubenswrapper[4839]: I0227 19:53:52.682211 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:53:52 crc kubenswrapper[4839]: I0227 19:53:52.682257 4839 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" Feb 27 19:53:52 crc kubenswrapper[4839]: I0227 19:53:52.682892 4839 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"324b7c3677fa3a46e6e785649ed814cb180ff19bad6941d855be7fc5a76021ea"} pod="openshift-machine-config-operator/machine-config-daemon-thb8n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 19:53:52 crc kubenswrapper[4839]: I0227 19:53:52.682960 4839 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" containerID="cri-o://324b7c3677fa3a46e6e785649ed814cb180ff19bad6941d855be7fc5a76021ea" gracePeriod=600 Feb 27 19:53:53 crc kubenswrapper[4839]: I0227 19:53:53.623696 4839 generic.go:334] "Generic (PLEG): container finished" podID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerID="324b7c3677fa3a46e6e785649ed814cb180ff19bad6941d855be7fc5a76021ea" exitCode=0 Feb 27 19:53:53 crc kubenswrapper[4839]: I0227 19:53:53.623820 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerDied","Data":"324b7c3677fa3a46e6e785649ed814cb180ff19bad6941d855be7fc5a76021ea"} Feb 27 19:53:53 crc kubenswrapper[4839]: I0227 19:53:53.624231 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" event={"ID":"f59a1827-952c-4b68-97ed-8d457c3cd3bf","Type":"ContainerStarted","Data":"83eea34b147969a266d4ecd87469a66951aae0f7dbcc470b99dd9f0f208f5d32"} Feb 27 19:53:53 crc kubenswrapper[4839]: I0227 19:53:53.624263 4839 scope.go:117] "RemoveContainer" containerID="cebc994c5b008539c690e93a6330de8ce7ecd9ea0602f50a6d17fdfe769f11e8" Feb 27 19:53:53 crc kubenswrapper[4839]: E0227 19:53:53.967648 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:53:56 crc kubenswrapper[4839]: E0227 19:53:56.967636 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:53:58 crc kubenswrapper[4839]: E0227 19:53:58.992310 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.154859 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537034-rs9g5"] Feb 27 19:54:00 crc kubenswrapper[4839]: E0227 19:54:00.155258 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="copy" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.155286 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="copy" Feb 27 19:54:00 crc kubenswrapper[4839]: E0227 19:54:00.155334 4839 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="gather" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.155351 4839 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="gather" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.155620 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="copy" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.155649 4839 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f1721c-699e-4d94-98aa-637641275f8b" containerName="gather" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.156400 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.166291 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537034-rs9g5"] Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.302787 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk2v2\" (UniqueName: \"kubernetes.io/projected/e88431ff-b79a-45f4-a09b-0e5a1c44171a-kube-api-access-vk2v2\") pod \"auto-csr-approver-29537034-rs9g5\" (UID: \"e88431ff-b79a-45f4-a09b-0e5a1c44171a\") " pod="openshift-infra/auto-csr-approver-29537034-rs9g5" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.405114 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk2v2\" (UniqueName: \"kubernetes.io/projected/e88431ff-b79a-45f4-a09b-0e5a1c44171a-kube-api-access-vk2v2\") pod \"auto-csr-approver-29537034-rs9g5\" (UID: \"e88431ff-b79a-45f4-a09b-0e5a1c44171a\") " pod="openshift-infra/auto-csr-approver-29537034-rs9g5" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.430716 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk2v2\" (UniqueName: \"kubernetes.io/projected/e88431ff-b79a-45f4-a09b-0e5a1c44171a-kube-api-access-vk2v2\") pod \"auto-csr-approver-29537034-rs9g5\" (UID: \"e88431ff-b79a-45f4-a09b-0e5a1c44171a\") " pod="openshift-infra/auto-csr-approver-29537034-rs9g5" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.485678 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" Feb 27 19:54:00 crc kubenswrapper[4839]: I0227 19:54:00.685021 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537034-rs9g5"] Feb 27 19:54:00 crc kubenswrapper[4839]: W0227 19:54:00.689459 4839 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode88431ff_b79a_45f4_a09b_0e5a1c44171a.slice/crio-3308328fcfb71f1c9f2609baeeea0923b9b629b6e366e547739b660ccbafe196 WatchSource:0}: Error finding container 3308328fcfb71f1c9f2609baeeea0923b9b629b6e366e547739b660ccbafe196: Status 404 returned error can't find the container with id 3308328fcfb71f1c9f2609baeeea0923b9b629b6e366e547739b660ccbafe196 Feb 27 19:54:01 crc kubenswrapper[4839]: I0227 19:54:01.691151 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" event={"ID":"e88431ff-b79a-45f4-a09b-0e5a1c44171a","Type":"ContainerStarted","Data":"3308328fcfb71f1c9f2609baeeea0923b9b629b6e366e547739b660ccbafe196"} Feb 27 19:54:01 crc kubenswrapper[4839]: E0227 19:54:01.935586 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:54:01 crc kubenswrapper[4839]: E0227 19:54:01.935992 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:54:01 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:54:01 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk2v2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537034-rs9g5_openshift-infra(e88431ff-b79a-45f4-a09b-0e5a1c44171a): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:54:01 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:54:01 crc kubenswrapper[4839]: E0227 19:54:01.937791 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:54:02 crc kubenswrapper[4839]: E0227 19:54:02.698182 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:54:03 crc kubenswrapper[4839]: E0227 19:54:03.968119 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:54:06 crc kubenswrapper[4839]: E0227 19:54:06.967715 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:54:11 crc kubenswrapper[4839]: E0227 19:54:11.966706 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:54:11 crc kubenswrapper[4839]: E0227 19:54:11.967281 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:54:16 crc kubenswrapper[4839]: E0227 19:54:16.884603 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:54:16 crc kubenswrapper[4839]: E0227 19:54:16.885623 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:54:16 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:54:16 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk2v2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537034-rs9g5_openshift-infra(e88431ff-b79a-45f4-a09b-0e5a1c44171a): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:54:16 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:54:16 crc kubenswrapper[4839]: E0227 19:54:16.886912 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:54:18 crc kubenswrapper[4839]: E0227 19:54:18.972489 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:54:19 crc kubenswrapper[4839]: E0227 19:54:19.966966 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:54:26 crc kubenswrapper[4839]: E0227 19:54:26.967780 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:54:26 crc kubenswrapper[4839]: E0227 19:54:26.968890 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:54:28 crc kubenswrapper[4839]: E0227 19:54:28.972130 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:54:31 crc kubenswrapper[4839]: E0227 19:54:31.968316 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:54:33 crc kubenswrapper[4839]: E0227 19:54:33.966654 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:54:39 crc kubenswrapper[4839]: E0227 19:54:39.967833 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:54:40 crc kubenswrapper[4839]: E0227 19:54:40.967220 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:54:41 crc kubenswrapper[4839]: E0227 19:54:41.603438 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:54:41 crc kubenswrapper[4839]: E0227 19:54:41.603907 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:54:41 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:54:41 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk2v2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537034-rs9g5_openshift-infra(e88431ff-b79a-45f4-a09b-0e5a1c44171a): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:54:41 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:54:41 crc kubenswrapper[4839]: E0227 19:54:41.605179 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:54:45 crc kubenswrapper[4839]: E0227 19:54:45.967421 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:54:48 crc kubenswrapper[4839]: E0227 19:54:48.906532 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:54:48 crc kubenswrapper[4839]: E0227 19:54:48.906716 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:54:48 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:54:48 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmf4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537032-wvf9m_openshift-infra(4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:54:48 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:54:48 crc kubenswrapper[4839]: E0227 19:54:48.907925 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:54:50 crc kubenswrapper[4839]: E0227 19:54:50.967891 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:54:53 crc kubenswrapper[4839]: E0227 19:54:53.967633 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:54:53 crc kubenswrapper[4839]: E0227 19:54:53.967908 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:54:56 crc kubenswrapper[4839]: E0227 19:54:56.967298 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:55:00 crc kubenswrapper[4839]: E0227 19:55:00.969252 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:55:01 crc kubenswrapper[4839]: E0227 19:55:01.967535 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:55:05 crc kubenswrapper[4839]: E0227 19:55:05.967260 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:55:05 crc kubenswrapper[4839]: E0227 19:55:05.967985 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:55:08 crc kubenswrapper[4839]: E0227 19:55:08.972645 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:55:11 crc kubenswrapper[4839]: E0227 19:55:11.967634 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:55:14 crc kubenswrapper[4839]: E0227 19:55:14.968446 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:55:16 crc kubenswrapper[4839]: E0227 19:55:16.967753 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:55:17 crc kubenswrapper[4839]: E0227 19:55:17.966905 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:55:19 crc kubenswrapper[4839]: E0227 19:55:19.967354 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:55:25 crc kubenswrapper[4839]: E0227 19:55:25.967335 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:55:28 crc kubenswrapper[4839]: E0227 19:55:28.973512 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:55:32 crc kubenswrapper[4839]: E0227 19:55:32.969679 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:55:33 crc kubenswrapper[4839]: E0227 19:55:33.967447 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:55:37 crc kubenswrapper[4839]: E0227 19:55:37.967805 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:55:43 crc kubenswrapper[4839]: E0227 19:55:43.968365 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:55:43 crc kubenswrapper[4839]: E0227 19:55:43.968431 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:55:45 crc kubenswrapper[4839]: E0227 19:55:45.967409 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:55:51 crc kubenswrapper[4839]: E0227 19:55:51.967013 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:55:54 crc kubenswrapper[4839]: E0227 19:55:54.967426 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:55:57 crc kubenswrapper[4839]: E0227 19:55:57.967926 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:55:59 crc kubenswrapper[4839]: E0227 19:55:59.967766 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.137439 4839 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537036-lckf6"] Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.138943 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.144714 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537036-lckf6"] Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.198857 4839 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gqv7\" (UniqueName: \"kubernetes.io/projected/0d459799-78a9-4aa1-ba12-77a144d5434e-kube-api-access-2gqv7\") pod \"auto-csr-approver-29537036-lckf6\" (UID: \"0d459799-78a9-4aa1-ba12-77a144d5434e\") " pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.300634 4839 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gqv7\" (UniqueName: \"kubernetes.io/projected/0d459799-78a9-4aa1-ba12-77a144d5434e-kube-api-access-2gqv7\") pod \"auto-csr-approver-29537036-lckf6\" (UID: \"0d459799-78a9-4aa1-ba12-77a144d5434e\") " pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.332542 4839 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gqv7\" (UniqueName: \"kubernetes.io/projected/0d459799-78a9-4aa1-ba12-77a144d5434e-kube-api-access-2gqv7\") pod \"auto-csr-approver-29537036-lckf6\" (UID: \"0d459799-78a9-4aa1-ba12-77a144d5434e\") " pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.461749 4839 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:00 crc kubenswrapper[4839]: I0227 19:56:00.639814 4839 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537036-lckf6"] Feb 27 19:56:01 crc kubenswrapper[4839]: I0227 19:56:01.489709 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537036-lckf6" event={"ID":"0d459799-78a9-4aa1-ba12-77a144d5434e","Type":"ContainerStarted","Data":"4a0800fb7ee0f1821974086f0d5b4154afb47fe1757911d0883a5bd0f20f1a68"} Feb 27 19:56:02 crc kubenswrapper[4839]: I0227 19:56:02.498225 4839 generic.go:334] "Generic (PLEG): container finished" podID="0d459799-78a9-4aa1-ba12-77a144d5434e" containerID="4dd554da1beb03a0b274aa570cbfca22ddc18b42b8534a1928017c727a11b02e" exitCode=0 Feb 27 19:56:02 crc kubenswrapper[4839]: I0227 19:56:02.498323 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537036-lckf6" event={"ID":"0d459799-78a9-4aa1-ba12-77a144d5434e","Type":"ContainerDied","Data":"4dd554da1beb03a0b274aa570cbfca22ddc18b42b8534a1928017c727a11b02e"} Feb 27 19:56:03 crc kubenswrapper[4839]: I0227 19:56:03.741963 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:03 crc kubenswrapper[4839]: I0227 19:56:03.849786 4839 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gqv7\" (UniqueName: \"kubernetes.io/projected/0d459799-78a9-4aa1-ba12-77a144d5434e-kube-api-access-2gqv7\") pod \"0d459799-78a9-4aa1-ba12-77a144d5434e\" (UID: \"0d459799-78a9-4aa1-ba12-77a144d5434e\") " Feb 27 19:56:03 crc kubenswrapper[4839]: I0227 19:56:03.854555 4839 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d459799-78a9-4aa1-ba12-77a144d5434e-kube-api-access-2gqv7" (OuterVolumeSpecName: "kube-api-access-2gqv7") pod "0d459799-78a9-4aa1-ba12-77a144d5434e" (UID: "0d459799-78a9-4aa1-ba12-77a144d5434e"). InnerVolumeSpecName "kube-api-access-2gqv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 19:56:03 crc kubenswrapper[4839]: I0227 19:56:03.951304 4839 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gqv7\" (UniqueName: \"kubernetes.io/projected/0d459799-78a9-4aa1-ba12-77a144d5434e-kube-api-access-2gqv7\") on node \"crc\" DevicePath \"\"" Feb 27 19:56:04 crc kubenswrapper[4839]: I0227 19:56:04.511220 4839 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537036-lckf6" event={"ID":"0d459799-78a9-4aa1-ba12-77a144d5434e","Type":"ContainerDied","Data":"4a0800fb7ee0f1821974086f0d5b4154afb47fe1757911d0883a5bd0f20f1a68"} Feb 27 19:56:04 crc kubenswrapper[4839]: I0227 19:56:04.511267 4839 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a0800fb7ee0f1821974086f0d5b4154afb47fe1757911d0883a5bd0f20f1a68" Feb 27 19:56:04 crc kubenswrapper[4839]: I0227 19:56:04.511272 4839 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537036-lckf6" Feb 27 19:56:04 crc kubenswrapper[4839]: I0227 19:56:04.812744 4839 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537026-67hjl"] Feb 27 19:56:04 crc kubenswrapper[4839]: I0227 19:56:04.816365 4839 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537026-67hjl"] Feb 27 19:56:04 crc kubenswrapper[4839]: E0227 19:56:04.968278 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:56:04 crc kubenswrapper[4839]: I0227 19:56:04.976965 4839 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b5f3e68-002d-4f8f-a93a-b4854e58fc5b" path="/var/lib/kubelet/pods/9b5f3e68-002d-4f8f-a93a-b4854e58fc5b/volumes" Feb 27 19:56:09 crc kubenswrapper[4839]: E0227 19:56:09.968305 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:56:10 crc kubenswrapper[4839]: E0227 19:56:10.968191 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:56:10 crc kubenswrapper[4839]: E0227 19:56:10.968583 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:56:18 crc kubenswrapper[4839]: E0227 19:56:18.970145 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:56:20 crc kubenswrapper[4839]: E0227 19:56:20.968350 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:56:21 crc kubenswrapper[4839]: E0227 19:56:21.967602 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:56:22 crc kubenswrapper[4839]: I0227 19:56:22.681423 4839 patch_prober.go:28] interesting pod/machine-config-daemon-thb8n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 19:56:22 crc kubenswrapper[4839]: I0227 19:56:22.681483 4839 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-thb8n" podUID="f59a1827-952c-4b68-97ed-8d457c3cd3bf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 19:56:24 crc kubenswrapper[4839]: E0227 19:56:24.967406 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:56:29 crc kubenswrapper[4839]: E0227 19:56:29.074167 4839 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 19:56:29 crc kubenswrapper[4839]: E0227 19:56:29.074556 4839 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 19:56:29 crc kubenswrapper[4839]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 19:56:29 crc kubenswrapper[4839]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk2v2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537034-rs9g5_openshift-infra(e88431ff-b79a-45f4-a09b-0e5a1c44171a): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 19:56:29 crc kubenswrapper[4839]: > logger="UnhandledError" Feb 27 19:56:29 crc kubenswrapper[4839]: E0227 19:56:29.075756 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:56:32 crc kubenswrapper[4839]: E0227 19:56:32.968544 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:56:35 crc kubenswrapper[4839]: E0227 19:56:35.967322 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:56:35 crc kubenswrapper[4839]: E0227 19:56:35.967685 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" Feb 27 19:56:39 crc kubenswrapper[4839]: E0227 19:56:39.966782 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mngj4" podUID="babba9f2-c281-4ad8-b697-9eb744a1960e" Feb 27 19:56:43 crc kubenswrapper[4839]: E0227 19:56:43.967021 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537034-rs9g5" podUID="e88431ff-b79a-45f4-a09b-0e5a1c44171a" Feb 27 19:56:44 crc kubenswrapper[4839]: E0227 19:56:44.966277 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537032-wvf9m" podUID="4b5affa9-4515-4f43-8d4e-dd8eb4d3ff97" Feb 27 19:56:46 crc kubenswrapper[4839]: E0227 19:56:46.967189 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7hm7c" podUID="02ff6284-1eb0-403a-8bfd-67184d2b3f5e" Feb 27 19:56:49 crc kubenswrapper[4839]: E0227 19:56:49.967078 4839 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vmj" podUID="827f2124-61ed-4c5e-acf7-21463af06877" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515150373415024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015150373416017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015150370327016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015150370327015460 5ustar corecore